[ 457.810316] env[61867]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61867) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 457.810649] env[61867]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61867) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 457.810802] env[61867]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61867) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 457.811142] env[61867]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 457.905627] env[61867]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61867) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 457.914892] env[61867]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.009s {{(pid=61867) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 458.516683] env[61867]: INFO nova.virt.driver [None req-f214986a-b298-400a-9cc2-6145ba06868b None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 458.588176] env[61867]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 458.588400] env[61867]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 458.588473] env[61867]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61867) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 461.660342] env[61867]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-1649bff8-bef2-48ee-aa10-2f0a2bca70f0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 461.676091] env[61867]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61867) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 461.676276] env[61867]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-93b3eb19-eda8-4b78-ad23-40a1e027ffed {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 461.700320] env[61867]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 342f8. [ 461.700464] env[61867]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.112s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 461.700953] env[61867]: INFO nova.virt.vmwareapi.driver [None req-f214986a-b298-400a-9cc2-6145ba06868b None None] VMware vCenter version: 7.0.3 [ 461.704245] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0eb87f-ecd4-4bc9-9cb8-4c6a8e7aab05 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 461.724762] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4359d9-87a8-4fde-9541-1c6a81939593 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 461.731452] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54dbf1c-cf8c-4478-af81-d695c8696f56 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 461.736762] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca71799-e975-43c5-897a-d50520347323 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 461.749670] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7b3e5e1-76f7-49bf-b182-c1ea70c0fc33 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 461.755287] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6195d102-5931-4ee1-9f45-b549bf74fc17 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 461.785544] env[61867]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-61531f64-20f2-4e85-8c65-60a30265b600 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 461.790378] env[61867]: DEBUG nova.virt.vmwareapi.driver [None req-f214986a-b298-400a-9cc2-6145ba06868b None None] Extension org.openstack.compute already exists. {{(pid=61867) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 461.792988] env[61867]: INFO nova.compute.provider_config [None req-f214986a-b298-400a-9cc2-6145ba06868b None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 462.295990] env[61867]: DEBUG nova.context [None req-f214986a-b298-400a-9cc2-6145ba06868b None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),8242b4f3-7c90-421b-8622-eddeb922dc96(cell1) {{(pid=61867) load_cells /opt/stack/nova/nova/context.py:464}} [ 462.302161] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 462.302161] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 462.302750] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 462.303427] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] Acquiring lock "8242b4f3-7c90-421b-8622-eddeb922dc96" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 462.303790] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] Lock "8242b4f3-7c90-421b-8622-eddeb922dc96" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 462.304931] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] Lock "8242b4f3-7c90-421b-8622-eddeb922dc96" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 462.325790] env[61867]: INFO dbcounter [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] Registered counter for database nova_cell0 [ 462.333953] env[61867]: INFO dbcounter [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] Registered counter for database nova_cell1 [ 462.337568] env[61867]: DEBUG oslo_db.sqlalchemy.engines [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61867) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 462.338064] env[61867]: DEBUG oslo_db.sqlalchemy.engines [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61867) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 462.342984] env[61867]: ERROR nova.db.main.api [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 462.342984] env[61867]: result = function(*args, **kwargs) [ 462.342984] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 462.342984] env[61867]: return func(*args, **kwargs) [ 462.342984] env[61867]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 462.342984] env[61867]: result = fn(*args, **kwargs) [ 462.342984] env[61867]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 462.342984] env[61867]: return f(*args, **kwargs) [ 462.342984] env[61867]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 462.342984] env[61867]: return db.service_get_minimum_version(context, binaries) [ 462.342984] env[61867]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 462.342984] env[61867]: _check_db_access() [ 462.342984] env[61867]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 462.342984] env[61867]: stacktrace = ''.join(traceback.format_stack()) [ 462.342984] env[61867]: [ 462.345449] env[61867]: ERROR nova.db.main.api [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 462.345449] env[61867]: result = function(*args, **kwargs) [ 462.345449] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 462.345449] env[61867]: return func(*args, **kwargs) [ 462.345449] env[61867]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 462.345449] env[61867]: result = fn(*args, **kwargs) [ 462.345449] env[61867]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 462.345449] env[61867]: return f(*args, **kwargs) [ 462.345449] env[61867]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 462.345449] env[61867]: return db.service_get_minimum_version(context, binaries) [ 462.345449] env[61867]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 462.345449] env[61867]: _check_db_access() [ 462.345449] env[61867]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 462.345449] env[61867]: stacktrace = ''.join(traceback.format_stack()) [ 462.345449] env[61867]: [ 462.345449] env[61867]: WARNING nova.objects.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] Failed to get minimum service version for cell 8242b4f3-7c90-421b-8622-eddeb922dc96 [ 462.345449] env[61867]: WARNING nova.objects.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 462.346270] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] Acquiring lock "singleton_lock" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 462.346570] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] Acquired lock "singleton_lock" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 462.346944] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] Releasing lock "singleton_lock" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 462.347431] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] Full set of CONF: {{(pid=61867) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 462.347727] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ******************************************************************************** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 462.347992] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] Configuration options gathered from: {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 462.348339] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 462.348670] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 462.348932] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ================================================================================ {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 462.349292] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] allow_resize_to_same_host = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.349602] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] arq_binding_timeout = 300 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.349863] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] backdoor_port = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.351028] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] backdoor_socket = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.351028] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] block_device_allocate_retries = 60 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.351028] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] block_device_allocate_retries_interval = 3 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.351028] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cert = self.pem {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.351028] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.351028] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] compute_monitors = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.351560] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] config_dir = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.351870] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] config_drive_format = iso9660 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.352142] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.352446] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] config_source = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.352754] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] console_host = devstack {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.353069] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] control_exchange = nova {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.353365] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cpu_allocation_ratio = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.353681] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] daemon = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.353980] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] debug = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.354313] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] default_access_ip_network_name = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.356016] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] default_availability_zone = nova {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.356016] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] default_ephemeral_format = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.356016] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] default_green_pool_size = 1000 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.356016] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.356016] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] default_schedule_zone = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.356231] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] disk_allocation_ratio = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.356231] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] enable_new_services = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.356231] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] enabled_apis = ['osapi_compute'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.356493] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] enabled_ssl_apis = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.356799] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] flat_injected = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360016] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] force_config_drive = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360016] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] force_raw_images = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360016] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] graceful_shutdown_timeout = 5 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360016] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] heal_instance_info_cache_interval = 60 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360016] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] host = cpu-1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360016] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360016] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] initial_disk_allocation_ratio = 1.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360258] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] initial_ram_allocation_ratio = 1.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360258] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360258] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] instance_build_timeout = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360258] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] instance_delete_interval = 300 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360258] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] instance_format = [instance: %(uuid)s] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360258] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] instance_name_template = instance-%08x {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360258] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] instance_usage_audit = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360434] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] instance_usage_audit_period = month {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360434] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360434] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] instances_path = /opt/stack/data/nova/instances {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360434] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] internal_service_availability_zone = internal {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.360839] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] key = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.361157] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] live_migration_retry_count = 30 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.361473] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] log_color = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.361777] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] log_config_append = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.362098] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.362395] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] log_dir = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.362735] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] log_file = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.362993] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] log_options = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.363311] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] log_rotate_interval = 1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.363618] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] log_rotate_interval_type = days {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.363914] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] log_rotation_type = none {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.364188] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.364443] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.364745] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.365057] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.365322] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.365651] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] long_rpc_timeout = 1800 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.365949] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] max_concurrent_builds = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.366279] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] max_concurrent_live_migrations = 1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.369076] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] max_concurrent_snapshots = 5 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.369076] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] max_local_block_devices = 3 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.369076] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] max_logfile_count = 30 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.369076] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] max_logfile_size_mb = 200 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.369076] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] maximum_instance_delete_attempts = 5 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.369076] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] metadata_listen = 0.0.0.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.369076] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] metadata_listen_port = 8775 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.369264] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] metadata_workers = 2 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.369264] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] migrate_max_retries = -1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.369264] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] mkisofs_cmd = genisoimage {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.369264] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] my_block_storage_ip = 10.180.1.21 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.369264] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] my_ip = 10.180.1.21 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.369264] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] network_allocate_retries = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.369264] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.369433] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] osapi_compute_listen = 0.0.0.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.369433] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] osapi_compute_listen_port = 8774 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.369734] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] osapi_compute_unique_server_name_scope = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.370057] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] osapi_compute_workers = 2 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.370360] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] password_length = 12 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.370660] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] periodic_enable = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.370960] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] periodic_fuzzy_delay = 60 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.371302] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] pointer_model = usbtablet {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.371636] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] preallocate_images = none {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.371942] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] publish_errors = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.372231] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] pybasedir = /opt/stack/nova {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.372518] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ram_allocation_ratio = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.372803] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] rate_limit_burst = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.373113] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] rate_limit_except_level = CRITICAL {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.373423] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] rate_limit_interval = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.373725] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] reboot_timeout = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.374036] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] reclaim_instance_interval = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.374328] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] record = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.374955] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] reimage_timeout_per_gb = 60 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.374955] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] report_interval = 120 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.375094] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] rescue_timeout = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.375172] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] reserved_host_cpus = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.375320] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] reserved_host_disk_mb = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.375505] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] reserved_host_memory_mb = 512 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.375664] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] reserved_huge_pages = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.375829] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] resize_confirm_window = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.375989] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] resize_fs_using_block_device = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.376160] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] resume_guests_state_on_host_boot = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.376327] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.376487] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] rpc_response_timeout = 60 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.376644] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] run_external_periodic_tasks = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.376805] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] running_deleted_instance_action = reap {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.376961] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] running_deleted_instance_poll_interval = 1800 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.377131] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] running_deleted_instance_timeout = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.377291] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] scheduler_instance_sync_interval = 120 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.377457] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] service_down_time = 720 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.377639] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] servicegroup_driver = db {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.377826] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] shell_completion = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.377985] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] shelved_offload_time = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.378159] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] shelved_poll_interval = 3600 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.378326] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] shutdown_timeout = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.378485] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] source_is_ipv6 = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.378643] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ssl_only = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.378908] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.379098] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] sync_power_state_interval = 600 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.379263] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] sync_power_state_pool_size = 1000 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.379430] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] syslog_log_facility = LOG_USER {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.379586] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] tempdir = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.379746] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] timeout_nbd = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.379913] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] transport_url = **** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.380085] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] update_resources_interval = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.380246] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] use_cow_images = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.380402] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] use_eventlog = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.380570] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] use_journal = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.380793] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] use_json = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.380970] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] use_rootwrap_daemon = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.381144] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] use_stderr = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.381305] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] use_syslog = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.381458] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vcpu_pin_set = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.381625] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vif_plugging_is_fatal = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.381790] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vif_plugging_timeout = 300 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.381953] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] virt_mkfs = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.382127] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] volume_usage_poll_interval = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.382288] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] watch_log_file = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.382453] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] web = /usr/share/spice-html5 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 462.382640] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.382807] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.382970] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.383161] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_concurrency.disable_process_locking = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.383901] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.384114] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.384295] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.384476] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.384652] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.384818] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.385076] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.auth_strategy = keystone {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.385184] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.compute_link_prefix = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.385360] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.385559] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.dhcp_domain = novalocal {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.385740] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.enable_instance_password = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.385908] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.glance_link_prefix = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.386086] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.386262] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.386441] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.instance_list_per_project_cells = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.386665] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.list_records_by_skipping_down_cells = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.386840] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.local_metadata_per_cell = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.387021] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.max_limit = 1000 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.387196] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.metadata_cache_expiration = 15 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.387373] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.neutron_default_tenant_id = default {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.387545] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.response_validation = warn {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.387718] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.use_neutron_default_nets = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.387886] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.388061] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.388261] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.388408] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.388582] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.vendordata_dynamic_targets = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.388748] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.vendordata_jsonfile_path = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.388931] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.389137] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.backend = dogpile.cache.memcached {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.389310] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.backend_argument = **** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.389498] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.config_prefix = cache.oslo {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.389686] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.dead_timeout = 60.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.389852] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.debug_cache_backend = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.390023] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.enable_retry_client = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.390191] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.enable_socket_keepalive = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.390362] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.enabled = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.390527] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.enforce_fips_mode = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.390692] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.expiration_time = 600 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.390854] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.hashclient_retry_attempts = 2 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.391028] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.hashclient_retry_delay = 1.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.391196] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.memcache_dead_retry = 300 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.391353] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.memcache_password = **** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.391530] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.391676] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.391840] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.memcache_pool_maxsize = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.392015] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.392186] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.memcache_sasl_enabled = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.392367] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.392554] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.memcache_socket_timeout = 1.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.392729] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.memcache_username = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.392903] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.proxies = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.393081] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.redis_db = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.393245] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.redis_password = **** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.393422] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.redis_sentinel_service_name = mymaster {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.393596] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.393763] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.redis_server = localhost:6379 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.393925] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.redis_socket_timeout = 1.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.394096] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.redis_username = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.394262] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.retry_attempts = 2 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.394424] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.retry_delay = 0.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.394585] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.socket_keepalive_count = 1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.394748] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.socket_keepalive_idle = 1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.394909] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.socket_keepalive_interval = 1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.395096] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.tls_allowed_ciphers = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.395255] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.tls_cafile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.395387] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.tls_certfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.395576] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.tls_enabled = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.395746] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cache.tls_keyfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.395920] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cinder.auth_section = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.396107] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cinder.auth_type = password {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.396272] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cinder.cafile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.396445] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cinder.catalog_info = volumev3::publicURL {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.396605] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cinder.certfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.396774] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cinder.collect_timing = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.396932] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cinder.cross_az_attach = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.397103] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cinder.debug = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.397264] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cinder.endpoint_template = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.397427] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cinder.http_retries = 3 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.397590] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cinder.insecure = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.397748] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cinder.keyfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.397917] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cinder.os_region_name = RegionOne {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.398090] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cinder.split_loggers = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.398253] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cinder.timeout = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.398421] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.398599] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] compute.cpu_dedicated_set = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.398772] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] compute.cpu_shared_set = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.398938] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] compute.image_type_exclude_list = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.399111] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.399278] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] compute.max_concurrent_disk_ops = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.399441] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] compute.max_disk_devices_to_attach = -1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.399602] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.399770] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.399932] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] compute.resource_provider_association_refresh = 300 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.400109] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.400273] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] compute.shutdown_retry_interval = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.400453] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.400634] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] conductor.workers = 2 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.400810] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] console.allowed_origins = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.400969] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] console.ssl_ciphers = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.401152] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] console.ssl_minimum_version = default {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.401324] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] consoleauth.enforce_session_timeout = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.401495] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] consoleauth.token_ttl = 600 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.401687] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.cafile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.401852] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.certfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.402022] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.collect_timing = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.402186] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.connect_retries = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.402342] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.connect_retry_delay = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.402499] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.endpoint_override = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.402658] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.insecure = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.402812] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.keyfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.402968] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.max_version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.403138] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.min_version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.403295] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.region_name = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.403452] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.retriable_status_codes = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.403610] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.service_name = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.403778] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.service_type = accelerator {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.403937] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.split_loggers = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.404104] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.status_code_retries = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.404262] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.status_code_retry_delay = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.404419] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.timeout = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.404612] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.404789] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] cyborg.version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.404987] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.backend = sqlalchemy {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.405217] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.connection = **** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.405317] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.connection_debug = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.405512] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.connection_parameters = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.405703] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.connection_recycle_time = 3600 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.405877] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.connection_trace = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.406049] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.db_inc_retry_interval = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.406218] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.db_max_retries = 20 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.406382] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.db_max_retry_interval = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.406547] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.db_retry_interval = 1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.406709] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.max_overflow = 50 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.406870] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.max_pool_size = 5 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.407081] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.max_retries = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.407324] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.407504] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.mysql_wsrep_sync_wait = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.407668] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.pool_timeout = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.407855] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.retry_interval = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.408040] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.slave_connection = **** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.408209] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.sqlite_synchronous = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.408371] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] database.use_db_reconnect = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.408549] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.backend = sqlalchemy {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.408747] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.connection = **** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.408914] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.connection_debug = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.409095] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.connection_parameters = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.409260] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.connection_recycle_time = 3600 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.409421] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.connection_trace = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.409579] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.db_inc_retry_interval = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.409740] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.db_max_retries = 20 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.409900] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.db_max_retry_interval = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.410095] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.db_retry_interval = 1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.410275] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.max_overflow = 50 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.410436] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.max_pool_size = 5 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.410597] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.max_retries = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.410767] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.410923] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.411090] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.pool_timeout = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.411254] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.retry_interval = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.411435] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.slave_connection = **** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.411621] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] api_database.sqlite_synchronous = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.411823] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] devices.enabled_mdev_types = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.412015] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.412197] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ephemeral_storage_encryption.default_format = luks {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.412359] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ephemeral_storage_encryption.enabled = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.412524] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.412696] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.api_servers = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.412857] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.cafile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.413026] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.certfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.413193] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.collect_timing = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.413352] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.connect_retries = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.413508] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.connect_retry_delay = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.413666] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.debug = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.413832] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.default_trusted_certificate_ids = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.413993] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.enable_certificate_validation = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.414166] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.enable_rbd_download = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.414324] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.endpoint_override = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.414515] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.insecure = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.414709] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.keyfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.414883] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.max_version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.415054] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.min_version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.415222] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.num_retries = 3 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.415393] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.rbd_ceph_conf = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.415598] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.rbd_connect_timeout = 5 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.415755] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.rbd_pool = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.415926] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.rbd_user = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.416098] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.region_name = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.416264] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.retriable_status_codes = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.416422] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.service_name = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.416592] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.service_type = image {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.416755] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.split_loggers = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.416927] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.status_code_retries = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.417091] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.status_code_retry_delay = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.417253] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.timeout = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.417442] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.417650] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.verify_glance_signatures = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.417836] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] glance.version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.418014] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] guestfs.debug = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.418189] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] mks.enabled = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.418550] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.418743] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] image_cache.manager_interval = 2400 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.418919] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] image_cache.precache_concurrency = 1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.419121] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] image_cache.remove_unused_base_images = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.419261] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.419428] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.419603] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] image_cache.subdirectory_name = _base {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.419782] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.api_max_retries = 60 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.419948] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.api_retry_interval = 2 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.420118] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.auth_section = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.420283] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.auth_type = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.420442] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.cafile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.420640] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.certfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.420828] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.collect_timing = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.420995] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.conductor_group = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.421170] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.connect_retries = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.421329] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.connect_retry_delay = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.421486] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.endpoint_override = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.421648] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.insecure = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.421804] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.keyfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.421959] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.max_version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.422132] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.min_version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.422294] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.peer_list = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.422449] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.region_name = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.422606] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.retriable_status_codes = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.422770] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.serial_console_state_timeout = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.422928] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.service_name = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.423107] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.service_type = baremetal {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.423682] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.shard = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.423682] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.split_loggers = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.423682] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.status_code_retries = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.423819] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.status_code_retry_delay = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.423981] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.timeout = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.424178] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.424340] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ironic.version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.424518] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.424689] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] key_manager.fixed_key = **** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.424867] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.425034] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican.barbican_api_version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.425199] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican.barbican_endpoint = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.425367] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican.barbican_endpoint_type = public {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.425549] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican.barbican_region_name = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.425728] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican.cafile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.425870] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican.certfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.426039] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican.collect_timing = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.426201] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican.insecure = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.426357] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican.keyfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.426561] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican.number_of_retries = 60 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.426746] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican.retry_delay = 1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.426938] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican.send_service_user_token = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.427116] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican.split_loggers = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.427277] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican.timeout = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.427436] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican.verify_ssl = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.427591] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican.verify_ssl_path = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.427755] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican_service_user.auth_section = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.427916] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican_service_user.auth_type = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.428083] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican_service_user.cafile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.428245] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican_service_user.certfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.428406] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican_service_user.collect_timing = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.428566] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican_service_user.insecure = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.428724] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican_service_user.keyfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.428883] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican_service_user.split_loggers = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.429050] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] barbican_service_user.timeout = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.429217] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vault.approle_role_id = **** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.429374] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vault.approle_secret_id = **** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.429583] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vault.kv_mountpoint = secret {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.429763] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vault.kv_path = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.429930] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vault.kv_version = 2 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.430102] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vault.namespace = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.430263] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vault.root_token_id = **** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.430419] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vault.ssl_ca_crt_file = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.430588] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vault.timeout = 60.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.430752] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vault.use_ssl = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.430920] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.431099] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.auth_section = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.431262] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.auth_type = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.431420] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.cafile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.431575] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.certfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.431738] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.collect_timing = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.431895] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.connect_retries = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.432059] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.connect_retry_delay = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.432220] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.endpoint_override = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.432379] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.insecure = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.432571] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.keyfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.432754] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.max_version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.432911] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.min_version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.433080] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.region_name = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.433242] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.retriable_status_codes = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.433397] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.service_name = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.433566] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.service_type = identity {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.433727] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.split_loggers = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.433883] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.status_code_retries = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.434048] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.status_code_retry_delay = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.434209] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.timeout = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.434388] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.434547] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] keystone.version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.434748] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.connection_uri = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.434908] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.cpu_mode = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.435084] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.cpu_model_extra_flags = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.435256] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.cpu_models = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.435425] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.cpu_power_governor_high = performance {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.435642] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.cpu_power_governor_low = powersave {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.435864] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.cpu_power_management = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.435999] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.436180] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.device_detach_attempts = 8 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.436344] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.device_detach_timeout = 20 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.436510] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.disk_cachemodes = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.436671] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.disk_prefix = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.436835] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.enabled_perf_events = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.436996] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.file_backed_memory = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.437174] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.gid_maps = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.437332] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.hw_disk_discard = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.437489] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.hw_machine_type = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.437659] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.images_rbd_ceph_conf = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.437828] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.437987] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.438165] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.images_rbd_glance_store_name = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.438330] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.images_rbd_pool = rbd {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.438498] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.images_type = default {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.438711] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.images_volume_group = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.438894] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.inject_key = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.439070] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.inject_partition = -2 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.439238] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.inject_password = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.439403] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.iscsi_iface = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.439563] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.iser_use_multipath = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.439727] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.live_migration_bandwidth = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.439888] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.440063] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.live_migration_downtime = 500 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.440231] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.440392] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.440555] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.live_migration_inbound_addr = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.440718] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.440884] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.live_migration_permit_post_copy = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.441054] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.live_migration_scheme = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.441233] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.live_migration_timeout_action = abort {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.441396] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.live_migration_tunnelled = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.441555] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.live_migration_uri = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.441766] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.live_migration_with_native_tls = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.441939] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.max_queues = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.442117] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.442364] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.442530] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.nfs_mount_options = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.442830] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.443021] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.443182] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.num_iser_scan_tries = 5 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.443343] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.num_memory_encrypted_guests = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.443507] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.443670] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.num_pcie_ports = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.443836] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.num_volume_scan_tries = 5 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.444008] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.pmem_namespaces = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.444173] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.quobyte_client_cfg = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.444455] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.444640] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.rbd_connect_timeout = 5 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.444842] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.445024] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.445190] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.rbd_secret_uuid = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.445349] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.rbd_user = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.445541] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.445722] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.remote_filesystem_transport = ssh {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.445882] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.rescue_image_id = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.446051] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.rescue_kernel_id = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.446208] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.rescue_ramdisk_id = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.446373] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.446536] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.rx_queue_size = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.446700] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.smbfs_mount_options = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.446971] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.447153] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.snapshot_compression = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.447314] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.snapshot_image_format = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.447530] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.447723] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.sparse_logical_volumes = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.447908] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.swtpm_enabled = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.448091] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.swtpm_group = tss {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.448259] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.swtpm_user = tss {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.448427] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.sysinfo_serial = unique {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.448583] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.tb_cache_size = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.448750] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.tx_queue_size = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.448917] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.uid_maps = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.449085] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.use_virtio_for_bridges = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.449256] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.virt_type = kvm {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.449420] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.volume_clear = zero {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.449582] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.volume_clear_size = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.449747] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.volume_use_multipath = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.449903] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.vzstorage_cache_path = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.450081] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.450248] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.vzstorage_mount_group = qemu {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.450411] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.vzstorage_mount_opts = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.450577] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.450889] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.451084] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.vzstorage_mount_user = stack {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.451257] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.451432] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.auth_section = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.451606] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.auth_type = password {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.451767] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.cafile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.451924] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.certfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.452095] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.collect_timing = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.452257] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.connect_retries = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.452413] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.connect_retry_delay = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.452580] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.default_floating_pool = public {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.452738] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.endpoint_override = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.452900] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.extension_sync_interval = 600 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.453068] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.http_retries = 3 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.453232] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.insecure = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.453390] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.keyfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.453548] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.max_version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.453729] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.453902] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.min_version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.454083] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.ovs_bridge = br-int {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.454252] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.physnets = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.454420] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.region_name = RegionOne {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.454579] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.retriable_status_codes = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.454748] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.service_metadata_proxy = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.454904] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.service_name = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.455081] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.service_type = network {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.455243] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.split_loggers = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.455401] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.status_code_retries = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.455587] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.status_code_retry_delay = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.455766] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.timeout = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.455927] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.456133] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] neutron.version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.456267] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] notifications.bdms_in_notifications = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.456493] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] notifications.default_level = INFO {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.456656] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] notifications.notification_format = unversioned {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.456827] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] notifications.notify_on_state_change = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.457035] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.457231] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] pci.alias = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.457402] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] pci.device_spec = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.457564] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] pci.report_in_placement = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.457773] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.auth_section = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.458070] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.auth_type = password {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.458320] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.458506] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.cafile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.458671] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.certfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.458837] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.collect_timing = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.458995] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.connect_retries = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.459169] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.connect_retry_delay = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.459326] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.default_domain_id = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.459483] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.default_domain_name = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.459641] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.domain_id = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.459799] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.domain_name = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.459953] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.endpoint_override = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.460124] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.insecure = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.460282] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.keyfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.460437] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.max_version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.460592] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.min_version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.460790] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.password = **** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.460971] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.project_domain_id = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.461154] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.project_domain_name = Default {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.461323] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.project_id = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.461525] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.project_name = service {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.461711] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.region_name = RegionOne {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.461972] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.retriable_status_codes = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.462245] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.service_name = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.462508] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.service_type = placement {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.462789] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.split_loggers = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.463075] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.status_code_retries = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.463359] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.status_code_retry_delay = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.463692] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.system_scope = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.463973] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.timeout = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.464267] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.trust_id = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.464522] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.user_domain_id = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.464731] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.user_domain_name = Default {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.464908] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.user_id = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.465124] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.username = nova {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.465322] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.465507] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] placement.version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.465714] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] quota.cores = 20 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.465899] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] quota.count_usage_from_placement = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.466084] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.466264] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] quota.injected_file_content_bytes = 10240 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.466431] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] quota.injected_file_path_length = 255 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.466601] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] quota.injected_files = 5 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.466789] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] quota.instances = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.466965] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] quota.key_pairs = 100 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.467148] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] quota.metadata_items = 128 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.467311] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] quota.ram = 51200 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.467474] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] quota.recheck_quota = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.467659] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] quota.server_group_members = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.467849] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] quota.server_groups = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.468101] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.468287] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.468456] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] scheduler.image_metadata_prefilter = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.468617] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.468811] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] scheduler.max_attempts = 3 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.468991] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] scheduler.max_placement_results = 1000 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.469170] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.469333] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] scheduler.query_placement_for_image_type_support = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.469495] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.469671] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] scheduler.workers = 2 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.469871] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.470068] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.470255] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.470426] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.470591] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.470760] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.470950] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.471172] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.471351] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.host_subset_size = 1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.471518] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.471678] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.471839] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.472056] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.isolated_hosts = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.472233] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.isolated_images = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.472398] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.472562] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.472731] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.472894] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.pci_in_placement = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.473065] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.473229] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.473396] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.473588] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.473770] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.473937] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.474153] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.track_instance_changes = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.474290] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.474442] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] metrics.required = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.474607] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] metrics.weight_multiplier = 1.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.474769] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.474931] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] metrics.weight_setting = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.475259] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.475438] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] serial_console.enabled = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.475631] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] serial_console.port_range = 10000:20000 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.475807] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.476029] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.476151] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] serial_console.serialproxy_port = 6083 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.476411] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] service_user.auth_section = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.476501] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] service_user.auth_type = password {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.476649] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] service_user.cafile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.476807] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] service_user.certfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.476966] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] service_user.collect_timing = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.477137] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] service_user.insecure = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.477292] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] service_user.keyfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.477461] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] service_user.send_service_user_token = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.477622] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] service_user.split_loggers = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.477794] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] service_user.timeout = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.477962] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] spice.agent_enabled = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.478137] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] spice.enabled = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.478445] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.478639] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.478814] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] spice.html5proxy_port = 6082 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.478978] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] spice.image_compression = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.479148] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] spice.jpeg_compression = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.479305] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] spice.playback_compression = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.479466] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] spice.require_secure = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.479635] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] spice.server_listen = 127.0.0.1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.479802] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.479957] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] spice.streaming_mode = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.480124] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] spice.zlib_compression = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.480292] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] upgrade_levels.baseapi = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.480456] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] upgrade_levels.compute = auto {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.480615] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] upgrade_levels.conductor = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.480770] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] upgrade_levels.scheduler = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.480933] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vendordata_dynamic_auth.auth_section = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.481106] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vendordata_dynamic_auth.auth_type = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.481274] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vendordata_dynamic_auth.cafile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.481417] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vendordata_dynamic_auth.certfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.481574] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.481735] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vendordata_dynamic_auth.insecure = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.481888] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vendordata_dynamic_auth.keyfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.482054] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.482214] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vendordata_dynamic_auth.timeout = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.482389] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.api_retry_count = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.482545] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.ca_file = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.482714] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.cache_prefix = devstack-image-cache {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.482878] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.cluster_name = testcl1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.483048] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.connection_pool_size = 10 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.483218] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.console_delay_seconds = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.483373] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.datastore_regex = ^datastore.* {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.483612] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.483791] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.host_password = **** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.483958] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.host_port = 443 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.484137] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.host_username = administrator@vsphere.local {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.484307] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.insecure = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.484466] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.integration_bridge = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.484626] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.maximum_objects = 100 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.484785] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.pbm_default_policy = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.484941] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.pbm_enabled = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.485108] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.pbm_wsdl_location = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.485275] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.485429] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.serial_port_proxy_uri = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.485637] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.serial_port_service_uri = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.485828] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.task_poll_interval = 0.5 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.486010] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.use_linked_clone = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.486185] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.vnc_keymap = en-us {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.486349] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.vnc_port = 5900 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.486572] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vmware.vnc_port_total = 10000 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.486694] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vnc.auth_schemes = ['none'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.486866] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vnc.enabled = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.487195] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.487382] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.487552] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vnc.novncproxy_port = 6080 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.487732] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vnc.server_listen = 127.0.0.1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.487903] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.488071] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vnc.vencrypt_ca_certs = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.488234] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vnc.vencrypt_client_cert = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.488390] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vnc.vencrypt_client_key = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.488565] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.488729] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.disable_deep_image_inspection = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.488887] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.489052] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.489212] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.489371] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.disable_rootwrap = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.489529] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.enable_numa_live_migration = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.489687] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.489842] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.489996] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.490163] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.libvirt_disable_apic = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.490318] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.490476] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.490636] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.490795] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.490949] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.491116] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.491274] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.491432] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.491588] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.491751] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.491933] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.492110] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] wsgi.client_socket_timeout = 900 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.492279] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] wsgi.default_pool_size = 1000 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.492445] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] wsgi.keep_alive = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.492612] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] wsgi.max_header_line = 16384 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.492771] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] wsgi.secure_proxy_ssl_header = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.492928] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] wsgi.ssl_ca_file = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.493095] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] wsgi.ssl_cert_file = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.493255] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] wsgi.ssl_key_file = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.493420] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] wsgi.tcp_keepidle = 600 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.493628] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.493805] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] zvm.ca_file = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.493965] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] zvm.cloud_connector_url = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.494292] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.494470] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] zvm.reachable_timeout = 300 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.494647] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_policy.enforce_new_defaults = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.495071] env[61867]: WARNING oslo_config.cfg [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 462.495258] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_policy.enforce_scope = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.495438] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_policy.policy_default_rule = default {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.495647] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.495832] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_policy.policy_file = policy.yaml {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.496016] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.496191] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.496352] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.496498] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.496761] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.496846] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.496981] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.497168] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] profiler.connection_string = messaging:// {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.497337] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] profiler.enabled = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.497504] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] profiler.es_doc_type = notification {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.497707] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] profiler.es_scroll_size = 10000 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.497898] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] profiler.es_scroll_time = 2m {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.498074] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] profiler.filter_error_trace = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.498247] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] profiler.hmac_keys = **** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.498415] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] profiler.sentinel_service_name = mymaster {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.498587] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] profiler.socket_timeout = 0.1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.498749] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] profiler.trace_requests = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.498908] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] profiler.trace_sqlalchemy = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.499099] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] profiler_jaeger.process_tags = {} {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.499259] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] profiler_jaeger.service_name_prefix = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.499420] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] profiler_otlp.service_name_prefix = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.499585] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] remote_debug.host = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.499742] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] remote_debug.port = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.499920] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.500093] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.500259] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.500421] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.500580] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.500741] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.500903] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.501075] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.501237] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.501404] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.501561] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.501731] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.501897] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.502076] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.502248] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.502411] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.502574] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.502746] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.502908] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.503077] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.503247] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.503414] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.503608] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.503788] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.503951] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.504126] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.504292] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.504451] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.504618] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.504784] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.ssl = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.504955] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.505138] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.505302] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.505478] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.505675] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.ssl_version = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.505841] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.506043] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.506215] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_notifications.retry = -1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.506478] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.506578] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_messaging_notifications.transport_url = **** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.506753] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.auth_section = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.506919] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.auth_type = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.507085] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.cafile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.507244] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.certfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.507410] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.collect_timing = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.507569] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.connect_retries = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.507728] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.connect_retry_delay = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.507884] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.endpoint_id = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.508045] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.endpoint_override = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.508209] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.insecure = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.508363] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.keyfile = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.508520] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.max_version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.508675] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.min_version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.508835] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.region_name = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.509068] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.retriable_status_codes = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.509245] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.service_name = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.509404] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.service_type = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.509565] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.split_loggers = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.509758] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.status_code_retries = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.509940] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.status_code_retry_delay = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.510114] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.timeout = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.510288] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.valid_interfaces = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.510430] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_limit.version = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.510886] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_reports.file_event_handler = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.510886] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.510886] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] oslo_reports.log_dir = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.511074] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.511237] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.511393] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.511559] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.511733] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.511896] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.512086] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.512251] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vif_plug_ovs_privileged.group = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.512412] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.512576] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.512741] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.512900] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] vif_plug_ovs_privileged.user = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.513081] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_vif_linux_bridge.flat_interface = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.513264] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.513436] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.513643] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.513824] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.513992] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.514172] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.514335] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.514512] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.514682] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_vif_ovs.isolate_vif = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.514851] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.515025] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.515195] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.515361] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_vif_ovs.ovsdb_interface = native {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.515546] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] os_vif_ovs.per_port_bridge = False {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.515732] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] privsep_osbrick.capabilities = [21] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.515893] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] privsep_osbrick.group = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.516060] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] privsep_osbrick.helper_command = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.516227] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.516391] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.516604] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] privsep_osbrick.user = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.516777] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.517049] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] nova_sys_admin.group = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.517190] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] nova_sys_admin.helper_command = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.517323] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.517489] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.517649] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] nova_sys_admin.user = None {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 462.517778] env[61867]: DEBUG oslo_service.service [None req-2f4bee0e-96ed-4ff4-9fb2-8f4b5b2e6eaa None None] ******************************************************************************** {{(pid=61867) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 462.518295] env[61867]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 463.022263] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Getting list of instances from cluster (obj){ [ 463.022263] env[61867]: value = "domain-c8" [ 463.022263] env[61867]: _type = "ClusterComputeResource" [ 463.022263] env[61867]: } {{(pid=61867) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 463.027676] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-828012e5-4483-4a33-96fe-2a8854a4387e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 463.037622] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Got total of 0 instances {{(pid=61867) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 463.037622] env[61867]: WARNING nova.virt.vmwareapi.driver [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 463.037622] env[61867]: INFO nova.virt.node [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Generated node identity 25720271-a549-4916-abe3-e5ed9b765889 [ 463.037622] env[61867]: INFO nova.virt.node [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Wrote node identity 25720271-a549-4916-abe3-e5ed9b765889 to /opt/stack/data/n-cpu-1/compute_id [ 463.541098] env[61867]: WARNING nova.compute.manager [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Compute nodes ['25720271-a549-4916-abe3-e5ed9b765889'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 464.547042] env[61867]: INFO nova.compute.manager [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 465.552571] env[61867]: WARNING nova.compute.manager [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 465.552939] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 465.553070] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 465.553212] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 465.553365] env[61867]: DEBUG nova.compute.resource_tracker [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61867) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 465.554315] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b1bd909-0174-4bfe-87ac-dde9d249e27b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 465.562841] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20accc98-82b7-482c-aa57-1e6f3bdcf602 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 465.576143] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a3aba0f-c738-47e3-92e6-993ad9605939 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 465.582628] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a3b4674-c45c-4515-b7cd-8d27ce3e8966 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 465.611204] env[61867]: DEBUG nova.compute.resource_tracker [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181513MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61867) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 465.611372] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 465.611529] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 466.114243] env[61867]: WARNING nova.compute.resource_tracker [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] No compute node record for cpu-1:25720271-a549-4916-abe3-e5ed9b765889: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 25720271-a549-4916-abe3-e5ed9b765889 could not be found. [ 466.618131] env[61867]: INFO nova.compute.resource_tracker [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 25720271-a549-4916-abe3-e5ed9b765889 [ 468.128916] env[61867]: DEBUG nova.compute.resource_tracker [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 468.129279] env[61867]: DEBUG nova.compute.resource_tracker [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 468.285263] env[61867]: INFO nova.scheduler.client.report [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] [req-9b631180-c344-4037-834e-c2d902581b5f] Created resource provider record via placement API for resource provider with UUID 25720271-a549-4916-abe3-e5ed9b765889 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 468.300641] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca4dda9-b5b9-4563-b0ff-438a1e9b2065 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.308046] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6b352b-28ff-4c51-969f-b8094134bbfb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.337114] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-befeaab6-27bb-4665-a6de-a4ff3aa2acba {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.343758] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32ced94-8f72-42c5-a0f3-785bee4bb6d7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.356238] env[61867]: DEBUG nova.compute.provider_tree [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 468.891184] env[61867]: DEBUG nova.scheduler.client.report [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Updated inventory for provider 25720271-a549-4916-abe3-e5ed9b765889 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 468.891429] env[61867]: DEBUG nova.compute.provider_tree [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Updating resource provider 25720271-a549-4916-abe3-e5ed9b765889 generation from 0 to 1 during operation: update_inventory {{(pid=61867) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 468.891571] env[61867]: DEBUG nova.compute.provider_tree [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 468.942193] env[61867]: DEBUG nova.compute.provider_tree [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Updating resource provider 25720271-a549-4916-abe3-e5ed9b765889 generation from 1 to 2 during operation: update_traits {{(pid=61867) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 469.447085] env[61867]: DEBUG nova.compute.resource_tracker [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61867) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 469.447436] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.836s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 469.447482] env[61867]: DEBUG nova.service [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Creating RPC server for service compute {{(pid=61867) start /opt/stack/nova/nova/service.py:186}} [ 469.461871] env[61867]: DEBUG nova.service [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] Join ServiceGroup membership for this service compute {{(pid=61867) start /opt/stack/nova/nova/service.py:203}} [ 469.462070] env[61867]: DEBUG nova.servicegroup.drivers.db [None req-ae017ed5-7137-4ef1-b478-2819c95b1661 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61867) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 504.457221] env[61867]: DEBUG oslo_concurrency.lockutils [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Acquiring lock "3afd10ad-17d4-47a7-b820-8f64622ea563" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 504.459159] env[61867]: DEBUG oslo_concurrency.lockutils [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Lock "3afd10ad-17d4-47a7-b820-8f64622ea563" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 504.964318] env[61867]: DEBUG nova.compute.manager [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 505.504554] env[61867]: DEBUG oslo_concurrency.lockutils [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 505.504929] env[61867]: DEBUG oslo_concurrency.lockutils [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 505.506736] env[61867]: INFO nova.compute.claims [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 506.580983] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab0fbeb6-43f4-49d8-a1ce-8fb1c8b007d2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.589699] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f04b8d20-828a-4573-8252-8c231573027e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.628314] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7352d25b-2a7b-4fcc-ab25-8f80d9643acf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.638408] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc000a41-4824-4c22-99c1-a3568450c3a2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.654488] env[61867]: DEBUG nova.compute.provider_tree [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 506.968848] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Acquiring lock "ef0db253-4aba-44ee-9986-b015f383dfc5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 506.969235] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Lock "ef0db253-4aba-44ee-9986-b015f383dfc5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 507.130680] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquiring lock "318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 507.130925] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Lock "318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 507.157789] env[61867]: DEBUG nova.scheduler.client.report [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 507.474176] env[61867]: DEBUG nova.compute.manager [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 507.634061] env[61867]: DEBUG nova.compute.manager [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 507.670025] env[61867]: DEBUG oslo_concurrency.lockutils [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.163s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 507.670025] env[61867]: DEBUG nova.compute.manager [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 507.902860] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Acquiring lock "510f3385-7dc0-4b73-b804-fa9e4f3ff995" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 507.902860] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Lock "510f3385-7dc0-4b73-b804-fa9e4f3ff995" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 508.007053] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 508.007053] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 508.008662] env[61867]: INFO nova.compute.claims [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 508.179754] env[61867]: DEBUG nova.compute.utils [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 508.184753] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 508.185080] env[61867]: DEBUG nova.compute.manager [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 508.186960] env[61867]: DEBUG nova.network.neutron [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 508.409318] env[61867]: DEBUG nova.compute.manager [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 508.693593] env[61867]: DEBUG nova.compute.manager [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 508.852609] env[61867]: DEBUG nova.policy [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4953960a87064f47b7af59cd84edee80', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b57dfbfc97842f7a55a7d6d28ffddd4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 508.949266] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 509.132927] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5fdcda7-56d8-4b61-8e5a-a587981e6d7f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.140812] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d988b2ed-bd61-4b54-a776-2319f77b31a1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.175826] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-177764f4-2cda-4e66-9bc0-43b80796e9ab {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.183977] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf140c94-3951-4bea-8bfc-e8a7cf786d7b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.206636] env[61867]: DEBUG nova.compute.provider_tree [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 509.696794] env[61867]: DEBUG nova.network.neutron [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Successfully created port: 1f1fb400-41bf-45d2-8b95-d9727ddf99ba {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 509.710516] env[61867]: DEBUG nova.compute.manager [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 509.723162] env[61867]: DEBUG nova.scheduler.client.report [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 509.756903] env[61867]: DEBUG nova.virt.hardware [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 509.756903] env[61867]: DEBUG nova.virt.hardware [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 509.756903] env[61867]: DEBUG nova.virt.hardware [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 509.757131] env[61867]: DEBUG nova.virt.hardware [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 509.758072] env[61867]: DEBUG nova.virt.hardware [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 509.758400] env[61867]: DEBUG nova.virt.hardware [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 509.758763] env[61867]: DEBUG nova.virt.hardware [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 509.759199] env[61867]: DEBUG nova.virt.hardware [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 509.759852] env[61867]: DEBUG nova.virt.hardware [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 509.760195] env[61867]: DEBUG nova.virt.hardware [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 509.760515] env[61867]: DEBUG nova.virt.hardware [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 509.761914] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c90b763-760a-40e3-9193-acd96f35f4f8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.775500] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e9f4b24-6563-43ee-93de-cc5415268444 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.793808] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb46a1d4-e36c-46b0-9b3d-4cd0aa96eeb3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.234394] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.226s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 510.234394] env[61867]: DEBUG nova.compute.manager [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 510.243039] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.056s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.243039] env[61867]: INFO nova.compute.claims [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 510.742686] env[61867]: DEBUG nova.compute.utils [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 510.745758] env[61867]: DEBUG nova.compute.manager [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 510.746093] env[61867]: DEBUG nova.network.neutron [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 510.929929] env[61867]: DEBUG nova.policy [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ea8a14b0ae374c4cad41618fd9c8ad91', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd49c191097244596b0f10ec9b53c95ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 511.248587] env[61867]: DEBUG nova.compute.manager [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 511.383204] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a42422cc-0fe0-4b85-bf1e-a34410a04426 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.395157] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b360d362-43b1-4cff-bb70-cb3194826984 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.432902] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db899fc-656b-4f53-9133-2bc27acc63cb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.441884] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44309a65-9810-4a84-879a-590154a33de2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.461465] env[61867]: DEBUG nova.compute.provider_tree [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 511.894603] env[61867]: ERROR nova.compute.manager [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1f1fb400-41bf-45d2-8b95-d9727ddf99ba, please check neutron logs for more information. [ 511.894603] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 511.894603] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 511.894603] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 511.894603] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 511.894603] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 511.894603] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 511.894603] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 511.894603] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 511.894603] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 511.894603] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 511.894603] env[61867]: ERROR nova.compute.manager raise self.value [ 511.894603] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 511.894603] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 511.894603] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 511.894603] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 511.898479] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 511.898479] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 511.898479] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1f1fb400-41bf-45d2-8b95-d9727ddf99ba, please check neutron logs for more information. [ 511.898479] env[61867]: ERROR nova.compute.manager [ 511.898479] env[61867]: Traceback (most recent call last): [ 511.898479] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 511.898479] env[61867]: listener.cb(fileno) [ 511.898479] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 511.898479] env[61867]: result = function(*args, **kwargs) [ 511.898479] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 511.898479] env[61867]: return func(*args, **kwargs) [ 511.898479] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 511.898479] env[61867]: raise e [ 511.898479] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 511.898479] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 511.898479] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 511.898479] env[61867]: created_port_ids = self._update_ports_for_instance( [ 511.898479] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 511.898479] env[61867]: with excutils.save_and_reraise_exception(): [ 511.898479] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 511.898479] env[61867]: self.force_reraise() [ 511.898479] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 511.898479] env[61867]: raise self.value [ 511.898479] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 511.898479] env[61867]: updated_port = self._update_port( [ 511.898479] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 511.898479] env[61867]: _ensure_no_port_binding_failure(port) [ 511.898479] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 511.898479] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 511.899454] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 1f1fb400-41bf-45d2-8b95-d9727ddf99ba, please check neutron logs for more information. [ 511.899454] env[61867]: Removing descriptor: 14 [ 511.899454] env[61867]: ERROR nova.compute.manager [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1f1fb400-41bf-45d2-8b95-d9727ddf99ba, please check neutron logs for more information. [ 511.899454] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Traceback (most recent call last): [ 511.899454] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 511.899454] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] yield resources [ 511.899454] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 511.899454] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] self.driver.spawn(context, instance, image_meta, [ 511.899454] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 511.899454] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] self._vmops.spawn(context, instance, image_meta, injected_files, [ 511.899454] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 511.899454] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] vm_ref = self.build_virtual_machine(instance, [ 511.900032] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 511.900032] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] vif_infos = vmwarevif.get_vif_info(self._session, [ 511.900032] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 511.900032] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] for vif in network_info: [ 511.900032] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 511.900032] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] return self._sync_wrapper(fn, *args, **kwargs) [ 511.900032] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 511.900032] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] self.wait() [ 511.900032] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 511.900032] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] self[:] = self._gt.wait() [ 511.900032] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 511.900032] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] return self._exit_event.wait() [ 511.900032] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 511.900360] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] result = hub.switch() [ 511.900360] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 511.900360] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] return self.greenlet.switch() [ 511.900360] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 511.900360] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] result = function(*args, **kwargs) [ 511.900360] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 511.900360] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] return func(*args, **kwargs) [ 511.900360] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 511.900360] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] raise e [ 511.900360] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 511.900360] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] nwinfo = self.network_api.allocate_for_instance( [ 511.900360] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 511.900360] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] created_port_ids = self._update_ports_for_instance( [ 511.900655] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 511.900655] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] with excutils.save_and_reraise_exception(): [ 511.900655] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 511.900655] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] self.force_reraise() [ 511.900655] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 511.900655] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] raise self.value [ 511.900655] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 511.900655] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] updated_port = self._update_port( [ 511.900655] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 511.900655] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] _ensure_no_port_binding_failure(port) [ 511.900655] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 511.900655] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] raise exception.PortBindingFailed(port_id=port['id']) [ 511.900933] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] nova.exception.PortBindingFailed: Binding failed for port 1f1fb400-41bf-45d2-8b95-d9727ddf99ba, please check neutron logs for more information. [ 511.900933] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] [ 511.900933] env[61867]: INFO nova.compute.manager [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Terminating instance [ 511.903869] env[61867]: DEBUG oslo_concurrency.lockutils [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Acquiring lock "refresh_cache-3afd10ad-17d4-47a7-b820-8f64622ea563" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 511.903869] env[61867]: DEBUG oslo_concurrency.lockutils [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Acquired lock "refresh_cache-3afd10ad-17d4-47a7-b820-8f64622ea563" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 511.903869] env[61867]: DEBUG nova.network.neutron [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 511.964395] env[61867]: DEBUG nova.scheduler.client.report [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 512.117470] env[61867]: DEBUG nova.network.neutron [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Successfully created port: 982825fe-91fc-472a-b53a-30d50370f33a {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 512.263215] env[61867]: DEBUG nova.compute.manager [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 512.307588] env[61867]: DEBUG nova.virt.hardware [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 512.307825] env[61867]: DEBUG nova.virt.hardware [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 512.307969] env[61867]: DEBUG nova.virt.hardware [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 512.308724] env[61867]: DEBUG nova.virt.hardware [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 512.308921] env[61867]: DEBUG nova.virt.hardware [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 512.309068] env[61867]: DEBUG nova.virt.hardware [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 512.310513] env[61867]: DEBUG nova.virt.hardware [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 512.310513] env[61867]: DEBUG nova.virt.hardware [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 512.310513] env[61867]: DEBUG nova.virt.hardware [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 512.311544] env[61867]: DEBUG nova.virt.hardware [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 512.311544] env[61867]: DEBUG nova.virt.hardware [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 512.312624] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5277d9a2-ef0c-4fb2-bc61-d979a103579c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.329619] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5ecc97-a661-4db3-8bd0-bee114a8950a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.470601] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.230s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 512.471861] env[61867]: DEBUG nova.compute.manager [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 512.475144] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.527s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 512.476684] env[61867]: INFO nova.compute.claims [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 512.605894] env[61867]: DEBUG nova.network.neutron [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 512.851789] env[61867]: DEBUG nova.network.neutron [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 512.985562] env[61867]: DEBUG nova.compute.utils [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 512.991184] env[61867]: DEBUG nova.compute.manager [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Not allocating networking since 'none' was specified. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 513.051013] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Acquiring lock "dad6d190-9e74-452b-bb76-6d62f283218e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.051359] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Lock "dad6d190-9e74-452b-bb76-6d62f283218e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.361230] env[61867]: DEBUG oslo_concurrency.lockutils [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Releasing lock "refresh_cache-3afd10ad-17d4-47a7-b820-8f64622ea563" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 513.361230] env[61867]: DEBUG nova.compute.manager [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 513.361385] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 513.361689] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a23491b8-3c86-4ecc-8e06-093aac6d6bb1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.374251] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eab1b0ef-6a87-4cdf-8439-56b76ce7e060 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.396113] env[61867]: DEBUG nova.compute.manager [req-ffaa12b7-9982-495a-886d-3613cbb43e0d req-a29676e6-8ff9-4db4-a790-e9d3ac85dcc0 service nova] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Received event network-changed-1f1fb400-41bf-45d2-8b95-d9727ddf99ba {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 513.396328] env[61867]: DEBUG nova.compute.manager [req-ffaa12b7-9982-495a-886d-3613cbb43e0d req-a29676e6-8ff9-4db4-a790-e9d3ac85dcc0 service nova] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Refreshing instance network info cache due to event network-changed-1f1fb400-41bf-45d2-8b95-d9727ddf99ba. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 513.396532] env[61867]: DEBUG oslo_concurrency.lockutils [req-ffaa12b7-9982-495a-886d-3613cbb43e0d req-a29676e6-8ff9-4db4-a790-e9d3ac85dcc0 service nova] Acquiring lock "refresh_cache-3afd10ad-17d4-47a7-b820-8f64622ea563" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 513.396800] env[61867]: DEBUG oslo_concurrency.lockutils [req-ffaa12b7-9982-495a-886d-3613cbb43e0d req-a29676e6-8ff9-4db4-a790-e9d3ac85dcc0 service nova] Acquired lock "refresh_cache-3afd10ad-17d4-47a7-b820-8f64622ea563" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 513.396895] env[61867]: DEBUG nova.network.neutron [req-ffaa12b7-9982-495a-886d-3613cbb43e0d req-a29676e6-8ff9-4db4-a790-e9d3ac85dcc0 service nova] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Refreshing network info cache for port 1f1fb400-41bf-45d2-8b95-d9727ddf99ba {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 513.417091] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3afd10ad-17d4-47a7-b820-8f64622ea563 could not be found. [ 513.417408] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 513.417966] env[61867]: INFO nova.compute.manager [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Took 0.06 seconds to destroy the instance on the hypervisor. [ 513.418347] env[61867]: DEBUG oslo.service.loopingcall [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 513.418622] env[61867]: DEBUG nova.compute.manager [-] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 513.418822] env[61867]: DEBUG nova.network.neutron [-] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 513.484333] env[61867]: DEBUG nova.network.neutron [-] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 513.492146] env[61867]: DEBUG nova.compute.manager [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 513.556641] env[61867]: DEBUG nova.compute.manager [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 513.615756] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e175c9c-d9b8-4fc0-95b1-c2c16f428b8f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.629768] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f889b339-5ce7-4b90-8b23-449fce719911 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.665508] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fdd1475-c4ce-4b75-b357-1f137f783693 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.676949] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d111f059-1455-45b2-be43-022e86b15091 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.694958] env[61867]: DEBUG nova.compute.provider_tree [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 513.930023] env[61867]: DEBUG nova.network.neutron [req-ffaa12b7-9982-495a-886d-3613cbb43e0d req-a29676e6-8ff9-4db4-a790-e9d3ac85dcc0 service nova] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 513.992344] env[61867]: DEBUG nova.network.neutron [-] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 514.081174] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.126195] env[61867]: DEBUG nova.network.neutron [req-ffaa12b7-9982-495a-886d-3613cbb43e0d req-a29676e6-8ff9-4db4-a790-e9d3ac85dcc0 service nova] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 514.198777] env[61867]: DEBUG nova.scheduler.client.report [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 514.495614] env[61867]: INFO nova.compute.manager [-] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Took 1.08 seconds to deallocate network for instance. [ 514.500104] env[61867]: DEBUG nova.compute.claims [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 514.500290] env[61867]: DEBUG oslo_concurrency.lockutils [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.502481] env[61867]: DEBUG nova.compute.manager [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 514.540487] env[61867]: DEBUG nova.virt.hardware [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 514.541016] env[61867]: DEBUG nova.virt.hardware [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 514.541016] env[61867]: DEBUG nova.virt.hardware [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 514.541125] env[61867]: DEBUG nova.virt.hardware [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 514.541350] env[61867]: DEBUG nova.virt.hardware [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 514.541350] env[61867]: DEBUG nova.virt.hardware [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 514.541954] env[61867]: DEBUG nova.virt.hardware [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 514.541954] env[61867]: DEBUG nova.virt.hardware [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 514.541954] env[61867]: DEBUG nova.virt.hardware [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 514.541954] env[61867]: DEBUG nova.virt.hardware [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 514.543733] env[61867]: DEBUG nova.virt.hardware [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 514.545345] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a211d1-12b5-4e2b-a298-aaa7870bc62d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.557601] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-619618be-17e7-4951-8afe-bf4b6480e8d1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.574017] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Instance VIF info [] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 514.585050] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 514.585846] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-efabb7de-8e51-4623-869f-dc6353d19ad6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.602696] env[61867]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 514.602911] env[61867]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61867) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 514.603257] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Folder already exists: OpenStack. Parent ref: group-v4. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 514.603486] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Creating folder: Project (34c9d429e04342379fea73177476877c). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 514.603942] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7bd18b83-f624-4f25-892d-8280e979bec7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.617880] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Created folder: Project (34c9d429e04342379fea73177476877c) in parent group-v274258. [ 514.618152] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Creating folder: Instances. Parent ref: group-v274262. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 514.618409] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-27f0a817-67a6-4259-955b-41085450ba3f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.630636] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Created folder: Instances in parent group-v274262. [ 514.631274] env[61867]: DEBUG oslo.service.loopingcall [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 514.633021] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 514.633021] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a679a3a2-0ff5-4169-a2b5-5c63faba1165 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.645807] env[61867]: DEBUG oslo_concurrency.lockutils [req-ffaa12b7-9982-495a-886d-3613cbb43e0d req-a29676e6-8ff9-4db4-a790-e9d3ac85dcc0 service nova] Releasing lock "refresh_cache-3afd10ad-17d4-47a7-b820-8f64622ea563" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 514.653683] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 514.653683] env[61867]: value = "task-1276284" [ 514.653683] env[61867]: _type = "Task" [ 514.653683] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 514.663362] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276284, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 514.709613] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.234s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 514.710166] env[61867]: DEBUG nova.compute.manager [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 514.713929] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.633s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.718280] env[61867]: INFO nova.compute.claims [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 514.953731] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Acquiring lock "433ce856-b04b-4808-8f05-201ad28e18d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.954266] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Lock "433ce856-b04b-4808-8f05-201ad28e18d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.958999] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Acquiring lock "b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.959226] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Lock "b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.168419] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276284, 'name': CreateVM_Task, 'duration_secs': 0.301664} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 515.168658] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 515.169600] env[61867]: DEBUG oslo_vmware.service [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b89872-a7b6-41f9-be81-c4c5887dc99d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.176711] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 515.178971] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 515.178971] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 515.178971] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ceaa640-b159-419f-9cf1-788612eb8611 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.183748] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 515.183748] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52303fb4-e6a3-6a8d-7928-573f81281b13" [ 515.183748] env[61867]: _type = "Task" [ 515.183748] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 515.192625] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52303fb4-e6a3-6a8d-7928-573f81281b13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 515.221913] env[61867]: DEBUG nova.compute.utils [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 515.228750] env[61867]: DEBUG nova.compute.manager [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 515.230549] env[61867]: DEBUG nova.network.neutron [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 515.379179] env[61867]: DEBUG nova.policy [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8db1f060f4044e5094c3341fe0125824', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c8ff16c03462415e991a5bd04db03782', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 515.460283] env[61867]: DEBUG nova.compute.manager [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 515.465676] env[61867]: DEBUG nova.compute.manager [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 515.697105] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 515.701433] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 515.701615] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 515.701868] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 515.702289] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 515.702565] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23900193-bdd1-49dd-87cd-8b3fbc600250 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.725025] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 515.725025] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 515.725025] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c6c789-0041-4cfe-8d62-1c52c0f9d5c2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.738342] env[61867]: DEBUG nova.compute.manager [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 515.742224] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33ef832f-5398-4310-aee2-8819cb791c99 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.752867] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 515.752867] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52dddd1c-6c57-8afd-5dd5-c40064882327" [ 515.752867] env[61867]: _type = "Task" [ 515.752867] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 515.767810] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Preparing fetch location {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 515.768078] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Creating directory with path [datastore2] vmware_temp/d6ce2bf6-92f4-4bdb-9f28-52caf329d618/4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 515.768345] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff4df4c1-8cb0-4a35-bdbf-e4947a500998 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.801721] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Created directory with path [datastore2] vmware_temp/d6ce2bf6-92f4-4bdb-9f28-52caf329d618/4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 515.801721] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Fetch image to [datastore2] vmware_temp/d6ce2bf6-92f4-4bdb-9f28-52caf329d618/4ca02567-c128-482f-b204-6ac166ac3160/tmp-sparse.vmdk {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 515.801721] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Downloading image file data 4ca02567-c128-482f-b204-6ac166ac3160 to [datastore2] vmware_temp/d6ce2bf6-92f4-4bdb-9f28-52caf329d618/4ca02567-c128-482f-b204-6ac166ac3160/tmp-sparse.vmdk on the data store datastore2 {{(pid=61867) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 515.806189] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04dfddf0-dd61-4bae-bc76-2ac1c74431f7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.820219] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3120e78d-4fdd-4d94-bfc6-23788b2cc389 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.836524] env[61867]: ERROR nova.compute.manager [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 982825fe-91fc-472a-b53a-30d50370f33a, please check neutron logs for more information. [ 515.836524] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 515.836524] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 515.836524] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 515.836524] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 515.836524] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 515.836524] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 515.836524] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 515.836524] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 515.836524] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 515.836524] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 515.836524] env[61867]: ERROR nova.compute.manager raise self.value [ 515.836524] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 515.836524] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 515.836524] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 515.836524] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 515.839189] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 515.839189] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 515.839189] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 982825fe-91fc-472a-b53a-30d50370f33a, please check neutron logs for more information. [ 515.839189] env[61867]: ERROR nova.compute.manager [ 515.839189] env[61867]: Traceback (most recent call last): [ 515.839189] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 515.839189] env[61867]: listener.cb(fileno) [ 515.839189] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 515.839189] env[61867]: result = function(*args, **kwargs) [ 515.839189] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 515.839189] env[61867]: return func(*args, **kwargs) [ 515.839189] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 515.839189] env[61867]: raise e [ 515.839189] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 515.839189] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 515.839189] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 515.839189] env[61867]: created_port_ids = self._update_ports_for_instance( [ 515.839189] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 515.839189] env[61867]: with excutils.save_and_reraise_exception(): [ 515.839189] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 515.839189] env[61867]: self.force_reraise() [ 515.839189] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 515.839189] env[61867]: raise self.value [ 515.839189] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 515.839189] env[61867]: updated_port = self._update_port( [ 515.839189] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 515.839189] env[61867]: _ensure_no_port_binding_failure(port) [ 515.839189] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 515.839189] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 515.839981] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 982825fe-91fc-472a-b53a-30d50370f33a, please check neutron logs for more information. [ 515.839981] env[61867]: Removing descriptor: 16 [ 515.839981] env[61867]: ERROR nova.compute.manager [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 982825fe-91fc-472a-b53a-30d50370f33a, please check neutron logs for more information. [ 515.839981] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Traceback (most recent call last): [ 515.839981] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 515.839981] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] yield resources [ 515.839981] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 515.839981] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] self.driver.spawn(context, instance, image_meta, [ 515.839981] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 515.839981] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 515.839981] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 515.839981] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] vm_ref = self.build_virtual_machine(instance, [ 515.840356] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 515.840356] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] vif_infos = vmwarevif.get_vif_info(self._session, [ 515.840356] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 515.840356] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] for vif in network_info: [ 515.840356] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 515.840356] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] return self._sync_wrapper(fn, *args, **kwargs) [ 515.840356] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 515.840356] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] self.wait() [ 515.840356] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 515.840356] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] self[:] = self._gt.wait() [ 515.840356] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 515.840356] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] return self._exit_event.wait() [ 515.840356] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 515.840676] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] result = hub.switch() [ 515.840676] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 515.840676] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] return self.greenlet.switch() [ 515.840676] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 515.840676] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] result = function(*args, **kwargs) [ 515.840676] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 515.840676] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] return func(*args, **kwargs) [ 515.840676] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 515.840676] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] raise e [ 515.840676] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 515.840676] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] nwinfo = self.network_api.allocate_for_instance( [ 515.840676] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 515.840676] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] created_port_ids = self._update_ports_for_instance( [ 515.842018] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 515.842018] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] with excutils.save_and_reraise_exception(): [ 515.842018] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 515.842018] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] self.force_reraise() [ 515.842018] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 515.842018] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] raise self.value [ 515.842018] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 515.842018] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] updated_port = self._update_port( [ 515.842018] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 515.842018] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] _ensure_no_port_binding_failure(port) [ 515.842018] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 515.842018] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] raise exception.PortBindingFailed(port_id=port['id']) [ 515.842480] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] nova.exception.PortBindingFailed: Binding failed for port 982825fe-91fc-472a-b53a-30d50370f33a, please check neutron logs for more information. [ 515.842480] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] [ 515.842480] env[61867]: INFO nova.compute.manager [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Terminating instance [ 515.843794] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Acquiring lock "refresh_cache-ef0db253-4aba-44ee-9986-b015f383dfc5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 515.843794] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Acquired lock "refresh_cache-ef0db253-4aba-44ee-9986-b015f383dfc5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 515.843794] env[61867]: DEBUG nova.network.neutron [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 515.855026] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b13c3cc-4d18-4994-a1a1-dafe483297a1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.905652] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af70411b-73f8-4cb0-9539-90f2a6166af9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.915080] env[61867]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-f68c58f6-c284-47c0-b182-e959b89d9e73 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.946748] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Downloading image file data 4ca02567-c128-482f-b204-6ac166ac3160 to the data store datastore2 {{(pid=61867) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 515.988428] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e76cb9-d8d4-40bb-9666-44636593bb6a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.001817] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-998a752a-e108-47e3-af5b-8de54f70000a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.011065] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.049236] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c3dacf-422d-4f83-84b6-43c0cd21aff7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.061549] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7655050c-cbee-4c31-8d20-69784101a7fb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.079071] env[61867]: DEBUG nova.compute.provider_tree [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 516.083747] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.094121] env[61867]: DEBUG nova.compute.manager [req-dfb95f74-d60e-4213-aa3b-38a620e5a10f req-1145ae21-0efe-47bf-bbe9-129b65e070c2 service nova] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Received event network-vif-deleted-1f1fb400-41bf-45d2-8b95-d9727ddf99ba {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 516.099792] env[61867]: DEBUG nova.compute.manager [req-4f8c1269-529b-4a5e-9f1f-8151aac6e912 req-7096b408-c47d-4c0a-8313-2a0aa66935ab service nova] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Received event network-changed-982825fe-91fc-472a-b53a-30d50370f33a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 516.100025] env[61867]: DEBUG nova.compute.manager [req-4f8c1269-529b-4a5e-9f1f-8151aac6e912 req-7096b408-c47d-4c0a-8313-2a0aa66935ab service nova] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Refreshing instance network info cache due to event network-changed-982825fe-91fc-472a-b53a-30d50370f33a. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 516.100833] env[61867]: DEBUG oslo_concurrency.lockutils [req-4f8c1269-529b-4a5e-9f1f-8151aac6e912 req-7096b408-c47d-4c0a-8313-2a0aa66935ab service nova] Acquiring lock "refresh_cache-ef0db253-4aba-44ee-9986-b015f383dfc5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 516.107012] env[61867]: DEBUG oslo_vmware.rw_handles [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/d6ce2bf6-92f4-4bdb-9f28-52caf329d618/4ca02567-c128-482f-b204-6ac166ac3160/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61867) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 516.587485] env[61867]: DEBUG nova.scheduler.client.report [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 516.639009] env[61867]: DEBUG nova.network.neutron [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 516.759049] env[61867]: DEBUG nova.compute.manager [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 516.798700] env[61867]: DEBUG nova.virt.hardware [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 516.799060] env[61867]: DEBUG nova.virt.hardware [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 516.802283] env[61867]: DEBUG nova.virt.hardware [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 516.802518] env[61867]: DEBUG nova.virt.hardware [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 516.802678] env[61867]: DEBUG nova.virt.hardware [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 516.802807] env[61867]: DEBUG nova.virt.hardware [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 516.803032] env[61867]: DEBUG nova.virt.hardware [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 516.803217] env[61867]: DEBUG nova.virt.hardware [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 516.803382] env[61867]: DEBUG nova.virt.hardware [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 516.803542] env[61867]: DEBUG nova.virt.hardware [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 516.803730] env[61867]: DEBUG nova.virt.hardware [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 516.804673] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-803959bd-9233-48cc-a563-6651a2ae27ba {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.823080] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-148a9c85-8acb-4f23-8d66-3fd20361ea1a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.870936] env[61867]: DEBUG nova.network.neutron [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 516.874456] env[61867]: DEBUG oslo_vmware.rw_handles [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Completed reading data from the image iterator. {{(pid=61867) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 516.874714] env[61867]: DEBUG oslo_vmware.rw_handles [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/d6ce2bf6-92f4-4bdb-9f28-52caf329d618/4ca02567-c128-482f-b204-6ac166ac3160/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61867) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 516.949131] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Downloaded image file data 4ca02567-c128-482f-b204-6ac166ac3160 to vmware_temp/d6ce2bf6-92f4-4bdb-9f28-52caf329d618/4ca02567-c128-482f-b204-6ac166ac3160/tmp-sparse.vmdk on the data store datastore2 {{(pid=61867) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 516.950974] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Caching image {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 516.951203] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Copying Virtual Disk [datastore2] vmware_temp/d6ce2bf6-92f4-4bdb-9f28-52caf329d618/4ca02567-c128-482f-b204-6ac166ac3160/tmp-sparse.vmdk to [datastore2] vmware_temp/d6ce2bf6-92f4-4bdb-9f28-52caf329d618/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 516.951484] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d2a9815c-34b9-462a-8271-726db147db30 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.958372] env[61867]: DEBUG nova.network.neutron [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Successfully created port: f1202e09-9714-4663-884f-7f8053ddc16a {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 516.965172] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 516.965172] env[61867]: value = "task-1276286" [ 516.965172] env[61867]: _type = "Task" [ 516.965172] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 516.975980] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276286, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 517.100294] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 517.101357] env[61867]: DEBUG nova.compute.manager [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 517.103684] env[61867]: DEBUG oslo_concurrency.lockutils [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 2.603s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.373511] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Releasing lock "refresh_cache-ef0db253-4aba-44ee-9986-b015f383dfc5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 517.373955] env[61867]: DEBUG nova.compute.manager [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 517.374197] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 517.374510] env[61867]: DEBUG oslo_concurrency.lockutils [req-4f8c1269-529b-4a5e-9f1f-8151aac6e912 req-7096b408-c47d-4c0a-8313-2a0aa66935ab service nova] Acquired lock "refresh_cache-ef0db253-4aba-44ee-9986-b015f383dfc5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 517.374687] env[61867]: DEBUG nova.network.neutron [req-4f8c1269-529b-4a5e-9f1f-8151aac6e912 req-7096b408-c47d-4c0a-8313-2a0aa66935ab service nova] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Refreshing network info cache for port 982825fe-91fc-472a-b53a-30d50370f33a {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 517.375746] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d4b142f7-430e-4027-a92a-b491717e5f8c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.391429] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec99ddd4-8543-4df4-900f-922cc2246e50 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.431850] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ef0db253-4aba-44ee-9986-b015f383dfc5 could not be found. [ 517.432100] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 517.432278] env[61867]: INFO nova.compute.manager [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Took 0.06 seconds to destroy the instance on the hypervisor. [ 517.434724] env[61867]: DEBUG oslo.service.loopingcall [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 517.435945] env[61867]: DEBUG nova.compute.manager [-] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 517.435945] env[61867]: DEBUG nova.network.neutron [-] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 517.477151] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276286, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 517.480300] env[61867]: DEBUG nova.network.neutron [-] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 517.611346] env[61867]: DEBUG nova.compute.utils [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 517.620630] env[61867]: DEBUG nova.compute.manager [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 517.620630] env[61867]: DEBUG nova.network.neutron [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 517.771238] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba07e28a-503d-427a-9803-d1bd25bd64ff {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.780788] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9942ef72-a000-4771-806a-820d1b43a29f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.822586] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd236fb-24ad-4c92-b5bb-ae38477f52a4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.836203] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e14dc0bc-4770-4494-9d15-f64ed914a4e0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.858387] env[61867]: DEBUG nova.compute.provider_tree [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 517.927092] env[61867]: DEBUG nova.policy [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cddd2dbd98a74914824d6a2c246b264e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '15aa3639fa9b4927b385a0786a8894af', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 517.963890] env[61867]: DEBUG nova.network.neutron [req-4f8c1269-529b-4a5e-9f1f-8151aac6e912 req-7096b408-c47d-4c0a-8313-2a0aa66935ab service nova] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 517.981845] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276286, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.694983} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 517.982193] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Copied Virtual Disk [datastore2] vmware_temp/d6ce2bf6-92f4-4bdb-9f28-52caf329d618/4ca02567-c128-482f-b204-6ac166ac3160/tmp-sparse.vmdk to [datastore2] vmware_temp/d6ce2bf6-92f4-4bdb-9f28-52caf329d618/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 517.982378] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Deleting the datastore file [datastore2] vmware_temp/d6ce2bf6-92f4-4bdb-9f28-52caf329d618/4ca02567-c128-482f-b204-6ac166ac3160/tmp-sparse.vmdk {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 517.982683] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f859237c-e394-4f79-ac8c-d2557589c84d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.986050] env[61867]: DEBUG nova.network.neutron [-] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 517.995737] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 517.995737] env[61867]: value = "task-1276288" [ 517.995737] env[61867]: _type = "Task" [ 517.995737] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 518.012157] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276288, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.117076] env[61867]: DEBUG nova.compute.manager [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 518.284217] env[61867]: DEBUG nova.network.neutron [req-4f8c1269-529b-4a5e-9f1f-8151aac6e912 req-7096b408-c47d-4c0a-8313-2a0aa66935ab service nova] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 518.361920] env[61867]: DEBUG nova.scheduler.client.report [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 518.463041] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 518.463375] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 518.463577] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Starting heal instance info cache {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 518.463714] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Rebuilding the list of instances to heal {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 518.493710] env[61867]: INFO nova.compute.manager [-] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Took 1.06 seconds to deallocate network for instance. [ 518.500715] env[61867]: DEBUG nova.compute.claims [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 518.501470] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.510265] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276288, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026789} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 518.510265] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 518.510265] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Moving file from [datastore2] vmware_temp/d6ce2bf6-92f4-4bdb-9f28-52caf329d618/4ca02567-c128-482f-b204-6ac166ac3160 to [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160. {{(pid=61867) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 518.510640] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-59c77e2a-472a-4720-9c5a-2eabe4d84a91 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.518860] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 518.518860] env[61867]: value = "task-1276289" [ 518.518860] env[61867]: _type = "Task" [ 518.518860] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 518.529934] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276289, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.788269] env[61867]: DEBUG oslo_concurrency.lockutils [req-4f8c1269-529b-4a5e-9f1f-8151aac6e912 req-7096b408-c47d-4c0a-8313-2a0aa66935ab service nova] Releasing lock "refresh_cache-ef0db253-4aba-44ee-9986-b015f383dfc5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 518.868934] env[61867]: DEBUG oslo_concurrency.lockutils [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.765s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 518.869710] env[61867]: ERROR nova.compute.manager [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1f1fb400-41bf-45d2-8b95-d9727ddf99ba, please check neutron logs for more information. [ 518.869710] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Traceback (most recent call last): [ 518.869710] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 518.869710] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] self.driver.spawn(context, instance, image_meta, [ 518.869710] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 518.869710] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] self._vmops.spawn(context, instance, image_meta, injected_files, [ 518.869710] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 518.869710] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] vm_ref = self.build_virtual_machine(instance, [ 518.869710] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 518.869710] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] vif_infos = vmwarevif.get_vif_info(self._session, [ 518.869710] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 518.871563] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] for vif in network_info: [ 518.871563] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 518.871563] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] return self._sync_wrapper(fn, *args, **kwargs) [ 518.871563] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 518.871563] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] self.wait() [ 518.871563] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 518.871563] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] self[:] = self._gt.wait() [ 518.871563] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 518.871563] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] return self._exit_event.wait() [ 518.871563] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 518.871563] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] result = hub.switch() [ 518.871563] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 518.871563] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] return self.greenlet.switch() [ 518.871966] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 518.871966] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] result = function(*args, **kwargs) [ 518.871966] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 518.871966] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] return func(*args, **kwargs) [ 518.871966] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 518.871966] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] raise e [ 518.871966] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 518.871966] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] nwinfo = self.network_api.allocate_for_instance( [ 518.871966] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 518.871966] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] created_port_ids = self._update_ports_for_instance( [ 518.871966] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 518.871966] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] with excutils.save_and_reraise_exception(): [ 518.871966] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 518.873480] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] self.force_reraise() [ 518.873480] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 518.873480] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] raise self.value [ 518.873480] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 518.873480] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] updated_port = self._update_port( [ 518.873480] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 518.873480] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] _ensure_no_port_binding_failure(port) [ 518.873480] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 518.873480] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] raise exception.PortBindingFailed(port_id=port['id']) [ 518.873480] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] nova.exception.PortBindingFailed: Binding failed for port 1f1fb400-41bf-45d2-8b95-d9727ddf99ba, please check neutron logs for more information. [ 518.873480] env[61867]: ERROR nova.compute.manager [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] [ 518.873874] env[61867]: DEBUG nova.compute.utils [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Binding failed for port 1f1fb400-41bf-45d2-8b95-d9727ddf99ba, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 518.873874] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.862s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.874252] env[61867]: INFO nova.compute.claims [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 518.881740] env[61867]: DEBUG nova.compute.manager [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Build of instance 3afd10ad-17d4-47a7-b820-8f64622ea563 was re-scheduled: Binding failed for port 1f1fb400-41bf-45d2-8b95-d9727ddf99ba, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 518.882415] env[61867]: DEBUG nova.compute.manager [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 518.882415] env[61867]: DEBUG oslo_concurrency.lockutils [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Acquiring lock "refresh_cache-3afd10ad-17d4-47a7-b820-8f64622ea563" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 518.882604] env[61867]: DEBUG oslo_concurrency.lockutils [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Acquired lock "refresh_cache-3afd10ad-17d4-47a7-b820-8f64622ea563" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 518.882642] env[61867]: DEBUG nova.network.neutron [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 518.969282] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 518.969460] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 518.969596] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 518.969723] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 518.969847] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Didn't find any instances for network info cache update. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 518.970388] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 518.970656] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 518.970859] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 518.971087] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 518.971519] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 518.971793] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 519.038269] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276289, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.153501} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 519.038386] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] File moved {{(pid=61867) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 519.038687] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Cleaning up location [datastore2] vmware_temp/d6ce2bf6-92f4-4bdb-9f28-52caf329d618 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 519.038863] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Deleting the datastore file [datastore2] vmware_temp/d6ce2bf6-92f4-4bdb-9f28-52caf329d618 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 519.039144] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4af54bbc-f1cc-4aa4-86d8-9350928b54fb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.049285] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 519.049285] env[61867]: value = "task-1276290" [ 519.049285] env[61867]: _type = "Task" [ 519.049285] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 519.063190] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276290, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 519.131587] env[61867]: DEBUG nova.compute.manager [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 519.172428] env[61867]: DEBUG nova.virt.hardware [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 519.173556] env[61867]: DEBUG nova.virt.hardware [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 519.173556] env[61867]: DEBUG nova.virt.hardware [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 519.173556] env[61867]: DEBUG nova.virt.hardware [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 519.173556] env[61867]: DEBUG nova.virt.hardware [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 519.176620] env[61867]: DEBUG nova.virt.hardware [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 519.176952] env[61867]: DEBUG nova.virt.hardware [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 519.177146] env[61867]: DEBUG nova.virt.hardware [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 519.177387] env[61867]: DEBUG nova.virt.hardware [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 519.177565] env[61867]: DEBUG nova.virt.hardware [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 519.177735] env[61867]: DEBUG nova.virt.hardware [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 519.179270] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab17d733-97ce-4c49-8732-66b534d43cbb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.192720] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce2b8ca-82d6-4317-a118-5111ac1c7785 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.338760] env[61867]: DEBUG nova.network.neutron [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Successfully created port: e0651f92-1ada-45ee-b225-f8bf6b977110 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 519.447967] env[61867]: DEBUG nova.network.neutron [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 519.479089] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Getting list of instances from cluster (obj){ [ 519.479089] env[61867]: value = "domain-c8" [ 519.479089] env[61867]: _type = "ClusterComputeResource" [ 519.479089] env[61867]: } {{(pid=61867) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 519.479089] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a411adae-eb5a-4728-b830-d7c5657367ef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.496104] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Got total of 1 instances {{(pid=61867) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 519.499242] env[61867]: WARNING nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] While synchronizing instance power states, found 5 instances in the database and 1 instances on the hypervisor. [ 519.499242] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Triggering sync for uuid ef0db253-4aba-44ee-9986-b015f383dfc5 {{(pid=61867) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 519.499242] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Triggering sync for uuid 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa {{(pid=61867) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 519.499242] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Triggering sync for uuid 510f3385-7dc0-4b73-b804-fa9e4f3ff995 {{(pid=61867) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 519.499242] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Triggering sync for uuid dad6d190-9e74-452b-bb76-6d62f283218e {{(pid=61867) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 519.499242] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Triggering sync for uuid b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7 {{(pid=61867) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 519.499242] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "ef0db253-4aba-44ee-9986-b015f383dfc5" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.499462] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.499462] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "510f3385-7dc0-4b73-b804-fa9e4f3ff995" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.499462] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "dad6d190-9e74-452b-bb76-6d62f283218e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.499462] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.499462] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 519.499608] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61867) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 519.504406] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 519.564701] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276290, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.0292} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 519.564701] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 519.564859] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1713199-172a-429d-a6af-e660be14764b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.572878] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 519.572878] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52905f99-8527-58a8-a43e-d4265225aa07" [ 519.572878] env[61867]: _type = "Task" [ 519.572878] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 519.585332] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52905f99-8527-58a8-a43e-d4265225aa07, 'name': SearchDatastore_Task, 'duration_secs': 0.008485} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 519.585332] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 519.585332] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa/318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 519.585332] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-75e46d72-2a1d-4ced-8b37-341fe0dd1d5f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.591658] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 519.591658] env[61867]: value = "task-1276291" [ 519.591658] env[61867]: _type = "Task" [ 519.591658] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 519.601547] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276291, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 519.840077] env[61867]: DEBUG nova.network.neutron [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 520.010725] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.053374] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82837e95-6587-4e2d-93fb-5e6367ad8c96 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.064569] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a0afef5-e24a-4815-a51a-fd5b30686792 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.102897] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ba97c2b-fe33-4d75-a4be-3a2c5d628d58 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.113806] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276291, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463321} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 520.115988] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa/318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 520.116276] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 520.116560] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7ab54727-bbcb-4cff-b150-5c321e5b35a6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.130292] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3361b6ec-bb85-4224-8062-3f082b79f7b7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.133515] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 520.133515] env[61867]: value = "task-1276293" [ 520.133515] env[61867]: _type = "Task" [ 520.133515] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 520.146272] env[61867]: DEBUG nova.compute.provider_tree [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 520.153414] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276293, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.246763] env[61867]: ERROR nova.compute.manager [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f1202e09-9714-4663-884f-7f8053ddc16a, please check neutron logs for more information. [ 520.246763] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 520.246763] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 520.246763] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 520.246763] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 520.246763] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 520.246763] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 520.246763] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 520.246763] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 520.246763] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 520.246763] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 520.246763] env[61867]: ERROR nova.compute.manager raise self.value [ 520.246763] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 520.246763] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 520.246763] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 520.246763] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 520.247264] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 520.247264] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 520.247264] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f1202e09-9714-4663-884f-7f8053ddc16a, please check neutron logs for more information. [ 520.247264] env[61867]: ERROR nova.compute.manager [ 520.247264] env[61867]: Traceback (most recent call last): [ 520.247264] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 520.247264] env[61867]: listener.cb(fileno) [ 520.247264] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 520.247264] env[61867]: result = function(*args, **kwargs) [ 520.247264] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 520.247264] env[61867]: return func(*args, **kwargs) [ 520.247264] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 520.247264] env[61867]: raise e [ 520.247264] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 520.247264] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 520.247264] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 520.247264] env[61867]: created_port_ids = self._update_ports_for_instance( [ 520.247264] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 520.247264] env[61867]: with excutils.save_and_reraise_exception(): [ 520.247264] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 520.247264] env[61867]: self.force_reraise() [ 520.247264] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 520.247264] env[61867]: raise self.value [ 520.247264] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 520.247264] env[61867]: updated_port = self._update_port( [ 520.247264] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 520.247264] env[61867]: _ensure_no_port_binding_failure(port) [ 520.247264] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 520.247264] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 520.248038] env[61867]: nova.exception.PortBindingFailed: Binding failed for port f1202e09-9714-4663-884f-7f8053ddc16a, please check neutron logs for more information. [ 520.248038] env[61867]: Removing descriptor: 14 [ 520.248038] env[61867]: ERROR nova.compute.manager [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f1202e09-9714-4663-884f-7f8053ddc16a, please check neutron logs for more information. [ 520.248038] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Traceback (most recent call last): [ 520.248038] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 520.248038] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] yield resources [ 520.248038] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 520.248038] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] self.driver.spawn(context, instance, image_meta, [ 520.248038] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 520.248038] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] self._vmops.spawn(context, instance, image_meta, injected_files, [ 520.248038] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 520.248038] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] vm_ref = self.build_virtual_machine(instance, [ 520.248477] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 520.248477] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] vif_infos = vmwarevif.get_vif_info(self._session, [ 520.248477] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 520.248477] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] for vif in network_info: [ 520.248477] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 520.248477] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] return self._sync_wrapper(fn, *args, **kwargs) [ 520.248477] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 520.248477] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] self.wait() [ 520.248477] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 520.248477] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] self[:] = self._gt.wait() [ 520.248477] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 520.248477] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] return self._exit_event.wait() [ 520.248477] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 520.248829] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] result = hub.switch() [ 520.248829] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 520.248829] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] return self.greenlet.switch() [ 520.248829] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 520.248829] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] result = function(*args, **kwargs) [ 520.248829] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 520.248829] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] return func(*args, **kwargs) [ 520.248829] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 520.248829] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] raise e [ 520.248829] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 520.248829] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] nwinfo = self.network_api.allocate_for_instance( [ 520.248829] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 520.248829] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] created_port_ids = self._update_ports_for_instance( [ 520.249198] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 520.249198] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] with excutils.save_and_reraise_exception(): [ 520.249198] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 520.249198] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] self.force_reraise() [ 520.249198] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 520.249198] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] raise self.value [ 520.249198] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 520.249198] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] updated_port = self._update_port( [ 520.249198] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 520.249198] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] _ensure_no_port_binding_failure(port) [ 520.249198] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 520.249198] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] raise exception.PortBindingFailed(port_id=port['id']) [ 520.249579] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] nova.exception.PortBindingFailed: Binding failed for port f1202e09-9714-4663-884f-7f8053ddc16a, please check neutron logs for more information. [ 520.249579] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] [ 520.249579] env[61867]: INFO nova.compute.manager [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Terminating instance [ 520.253357] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Acquiring lock "refresh_cache-510f3385-7dc0-4b73-b804-fa9e4f3ff995" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 520.253357] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Acquired lock "refresh_cache-510f3385-7dc0-4b73-b804-fa9e4f3ff995" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 520.253357] env[61867]: DEBUG nova.network.neutron [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 520.342330] env[61867]: DEBUG oslo_concurrency.lockutils [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Releasing lock "refresh_cache-3afd10ad-17d4-47a7-b820-8f64622ea563" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 520.343152] env[61867]: DEBUG nova.compute.manager [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 520.344435] env[61867]: DEBUG nova.compute.manager [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 520.344435] env[61867]: DEBUG nova.network.neutron [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 520.377101] env[61867]: DEBUG nova.network.neutron [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 520.650043] env[61867]: DEBUG nova.scheduler.client.report [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 520.653170] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276293, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.655229] env[61867]: DEBUG nova.compute.manager [req-27c505d5-914a-4c58-a89f-578078f4f55b req-f038de2a-78c0-430b-9d75-ece2599704ab service nova] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Received event network-vif-deleted-982825fe-91fc-472a-b53a-30d50370f33a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 520.781821] env[61867]: DEBUG nova.network.neutron [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 520.838690] env[61867]: DEBUG nova.network.neutron [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 520.880513] env[61867]: DEBUG nova.network.neutron [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 521.157667] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.285s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 521.157990] env[61867]: DEBUG nova.compute.manager [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 521.161091] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276293, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.997683} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.161888] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.078s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.163072] env[61867]: INFO nova.compute.claims [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 521.170546] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 521.170546] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e51ec5a-a111-46eb-afea-1b9dc432aa25 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.200630] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa/318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 521.200916] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45822406-3e9f-49c6-9235-38984e3a5616 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.222966] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 521.222966] env[61867]: value = "task-1276294" [ 521.222966] env[61867]: _type = "Task" [ 521.222966] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.237570] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276294, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.341992] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Releasing lock "refresh_cache-510f3385-7dc0-4b73-b804-fa9e4f3ff995" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 521.342120] env[61867]: DEBUG nova.compute.manager [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 521.342911] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 521.343208] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-53cb426b-66a6-4391-8a2d-7a89f8c43289 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.354729] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3bbd5eb-5d95-4fbc-8690-37ddf06106dc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.388578] env[61867]: INFO nova.compute.manager [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] [instance: 3afd10ad-17d4-47a7-b820-8f64622ea563] Took 1.04 seconds to deallocate network for instance. [ 521.391887] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 510f3385-7dc0-4b73-b804-fa9e4f3ff995 could not be found. [ 521.392423] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 521.392423] env[61867]: INFO nova.compute.manager [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Took 0.05 seconds to destroy the instance on the hypervisor. [ 521.392626] env[61867]: DEBUG oslo.service.loopingcall [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 521.393249] env[61867]: DEBUG nova.compute.manager [-] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 521.393377] env[61867]: DEBUG nova.network.neutron [-] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 521.430971] env[61867]: DEBUG nova.network.neutron [-] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 521.670198] env[61867]: DEBUG nova.compute.utils [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 521.674030] env[61867]: DEBUG nova.compute.manager [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 521.674030] env[61867]: DEBUG nova.network.neutron [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 521.736013] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276294, 'name': ReconfigVM_Task, 'duration_secs': 0.307002} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.736528] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa/318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 521.737893] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ccd981fe-e777-48a3-87a2-13bdc91d9806 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.745738] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 521.745738] env[61867]: value = "task-1276296" [ 521.745738] env[61867]: _type = "Task" [ 521.745738] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.757222] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276296, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.810758] env[61867]: DEBUG nova.policy [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cddd2dbd98a74914824d6a2c246b264e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '15aa3639fa9b4927b385a0786a8894af', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 521.933943] env[61867]: DEBUG nova.network.neutron [-] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 521.942658] env[61867]: DEBUG nova.compute.manager [req-54046a98-0059-4a00-ba69-8ab91150a948 req-89b6514d-624f-42f0-8a6e-aee93b43db99 service nova] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Received event network-changed-f1202e09-9714-4663-884f-7f8053ddc16a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 521.943054] env[61867]: DEBUG nova.compute.manager [req-54046a98-0059-4a00-ba69-8ab91150a948 req-89b6514d-624f-42f0-8a6e-aee93b43db99 service nova] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Refreshing instance network info cache due to event network-changed-f1202e09-9714-4663-884f-7f8053ddc16a. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 521.945063] env[61867]: DEBUG oslo_concurrency.lockutils [req-54046a98-0059-4a00-ba69-8ab91150a948 req-89b6514d-624f-42f0-8a6e-aee93b43db99 service nova] Acquiring lock "refresh_cache-510f3385-7dc0-4b73-b804-fa9e4f3ff995" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.945063] env[61867]: DEBUG oslo_concurrency.lockutils [req-54046a98-0059-4a00-ba69-8ab91150a948 req-89b6514d-624f-42f0-8a6e-aee93b43db99 service nova] Acquired lock "refresh_cache-510f3385-7dc0-4b73-b804-fa9e4f3ff995" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.945063] env[61867]: DEBUG nova.network.neutron [req-54046a98-0059-4a00-ba69-8ab91150a948 req-89b6514d-624f-42f0-8a6e-aee93b43db99 service nova] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Refreshing network info cache for port f1202e09-9714-4663-884f-7f8053ddc16a {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 522.180219] env[61867]: DEBUG nova.compute.manager [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 522.260080] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276296, 'name': Rename_Task, 'duration_secs': 0.156407} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.263279] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 522.263896] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0dea867-60c8-470a-9d8e-ac35417e4abd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.273682] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 522.273682] env[61867]: value = "task-1276297" [ 522.273682] env[61867]: _type = "Task" [ 522.273682] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.285034] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276297, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.425774] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f93543-11b7-4d19-9489-aa4ed1eaa2ea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.436392] env[61867]: INFO nova.scheduler.client.report [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Deleted allocations for instance 3afd10ad-17d4-47a7-b820-8f64622ea563 [ 522.443531] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81cd6a64-ed96-458f-bb39-4b0a3c9e699f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.447674] env[61867]: INFO nova.compute.manager [-] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Took 1.05 seconds to deallocate network for instance. [ 522.453220] env[61867]: DEBUG nova.compute.claims [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 522.453647] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.491251] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Acquiring lock "ac5df149-83db-481a-a240-e5cf08fdbff9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.491602] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Lock "ac5df149-83db-481a-a240-e5cf08fdbff9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.494714] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f33e8b-a1e2-41e3-956c-898533a7636d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.506364] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c64e0e51-f754-4348-b09a-dd26c1a4bf6a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.524980] env[61867]: DEBUG nova.compute.provider_tree [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 522.534966] env[61867]: DEBUG nova.network.neutron [req-54046a98-0059-4a00-ba69-8ab91150a948 req-89b6514d-624f-42f0-8a6e-aee93b43db99 service nova] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 522.788644] env[61867]: DEBUG oslo_vmware.api [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276297, 'name': PowerOnVM_Task, 'duration_secs': 0.501617} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.789222] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 522.789557] env[61867]: INFO nova.compute.manager [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Took 8.29 seconds to spawn the instance on the hypervisor. [ 522.789929] env[61867]: DEBUG nova.compute.manager [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 522.791350] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e85bcc-2b06-40b7-85bc-dc22989724f0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.855019] env[61867]: DEBUG nova.network.neutron [req-54046a98-0059-4a00-ba69-8ab91150a948 req-89b6514d-624f-42f0-8a6e-aee93b43db99 service nova] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 522.939477] env[61867]: DEBUG nova.network.neutron [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Successfully created port: f757b37f-f7bf-48b8-be0d-0ad5e3eadd6b {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 522.953944] env[61867]: DEBUG oslo_concurrency.lockutils [None req-074a4ce4-5463-4ea6-be27-00a785ad012f tempest-FloatingIPsAssociationNegativeTestJSON-1254672853 tempest-FloatingIPsAssociationNegativeTestJSON-1254672853-project-member] Lock "3afd10ad-17d4-47a7-b820-8f64622ea563" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.496s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.995470] env[61867]: DEBUG nova.compute.manager [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 523.028788] env[61867]: DEBUG nova.scheduler.client.report [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 523.194916] env[61867]: DEBUG nova.compute.manager [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 523.220144] env[61867]: DEBUG nova.virt.hardware [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 523.220656] env[61867]: DEBUG nova.virt.hardware [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 523.220897] env[61867]: DEBUG nova.virt.hardware [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 523.221156] env[61867]: DEBUG nova.virt.hardware [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 523.221341] env[61867]: DEBUG nova.virt.hardware [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 523.221460] env[61867]: DEBUG nova.virt.hardware [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 523.221703] env[61867]: DEBUG nova.virt.hardware [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 523.221971] env[61867]: DEBUG nova.virt.hardware [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 523.222190] env[61867]: DEBUG nova.virt.hardware [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 523.222420] env[61867]: DEBUG nova.virt.hardware [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 523.223087] env[61867]: DEBUG nova.virt.hardware [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 523.225125] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5546aa46-dfc8-4067-968a-d005e2b4cca7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.235508] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6d9d565-b760-46c7-a1b4-4a257fc40a84 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.316047] env[61867]: INFO nova.compute.manager [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Took 15.17 seconds to build instance. [ 523.358869] env[61867]: DEBUG oslo_concurrency.lockutils [req-54046a98-0059-4a00-ba69-8ab91150a948 req-89b6514d-624f-42f0-8a6e-aee93b43db99 service nova] Releasing lock "refresh_cache-510f3385-7dc0-4b73-b804-fa9e4f3ff995" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.524837] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.534399] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 523.534879] env[61867]: DEBUG nova.compute.manager [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 523.537476] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 5.036s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.819461] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e04461d0-4b2d-44c8-88eb-d0ba856fe227 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Lock "318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.688s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 523.819461] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 4.321s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.819569] env[61867]: INFO nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] During sync_power_state the instance has a pending task (spawning). Skip. [ 523.819757] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 524.050380] env[61867]: DEBUG nova.compute.utils [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 524.055064] env[61867]: DEBUG nova.compute.manager [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 524.055064] env[61867]: DEBUG nova.network.neutron [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 524.107995] env[61867]: ERROR nova.compute.manager [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e0651f92-1ada-45ee-b225-f8bf6b977110, please check neutron logs for more information. [ 524.107995] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 524.107995] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.107995] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 524.107995] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 524.107995] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 524.107995] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 524.107995] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 524.107995] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.107995] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 524.107995] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.107995] env[61867]: ERROR nova.compute.manager raise self.value [ 524.107995] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 524.107995] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 524.107995] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.107995] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 524.108481] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.108481] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 524.108481] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e0651f92-1ada-45ee-b225-f8bf6b977110, please check neutron logs for more information. [ 524.108481] env[61867]: ERROR nova.compute.manager [ 524.108481] env[61867]: Traceback (most recent call last): [ 524.108481] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 524.108481] env[61867]: listener.cb(fileno) [ 524.108481] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.108481] env[61867]: result = function(*args, **kwargs) [ 524.108481] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 524.108481] env[61867]: return func(*args, **kwargs) [ 524.108481] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 524.108481] env[61867]: raise e [ 524.108481] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.108481] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 524.108481] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 524.108481] env[61867]: created_port_ids = self._update_ports_for_instance( [ 524.108481] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 524.108481] env[61867]: with excutils.save_and_reraise_exception(): [ 524.108481] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.108481] env[61867]: self.force_reraise() [ 524.108481] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.108481] env[61867]: raise self.value [ 524.108481] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 524.108481] env[61867]: updated_port = self._update_port( [ 524.108481] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.108481] env[61867]: _ensure_no_port_binding_failure(port) [ 524.108481] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.108481] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 524.109246] env[61867]: nova.exception.PortBindingFailed: Binding failed for port e0651f92-1ada-45ee-b225-f8bf6b977110, please check neutron logs for more information. [ 524.109246] env[61867]: Removing descriptor: 16 [ 524.109246] env[61867]: ERROR nova.compute.manager [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e0651f92-1ada-45ee-b225-f8bf6b977110, please check neutron logs for more information. [ 524.109246] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Traceback (most recent call last): [ 524.109246] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 524.109246] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] yield resources [ 524.109246] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 524.109246] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] self.driver.spawn(context, instance, image_meta, [ 524.109246] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 524.109246] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 524.109246] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 524.109246] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] vm_ref = self.build_virtual_machine(instance, [ 524.109561] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 524.109561] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] vif_infos = vmwarevif.get_vif_info(self._session, [ 524.109561] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 524.109561] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] for vif in network_info: [ 524.109561] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 524.109561] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] return self._sync_wrapper(fn, *args, **kwargs) [ 524.109561] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 524.109561] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] self.wait() [ 524.109561] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 524.109561] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] self[:] = self._gt.wait() [ 524.109561] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 524.109561] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] return self._exit_event.wait() [ 524.109561] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 524.109935] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] result = hub.switch() [ 524.109935] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 524.109935] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] return self.greenlet.switch() [ 524.109935] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.109935] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] result = function(*args, **kwargs) [ 524.109935] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 524.109935] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] return func(*args, **kwargs) [ 524.109935] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 524.109935] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] raise e [ 524.109935] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.109935] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] nwinfo = self.network_api.allocate_for_instance( [ 524.109935] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 524.109935] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] created_port_ids = self._update_ports_for_instance( [ 524.110303] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 524.110303] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] with excutils.save_and_reraise_exception(): [ 524.110303] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.110303] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] self.force_reraise() [ 524.110303] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.110303] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] raise self.value [ 524.110303] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 524.110303] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] updated_port = self._update_port( [ 524.110303] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.110303] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] _ensure_no_port_binding_failure(port) [ 524.110303] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.110303] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] raise exception.PortBindingFailed(port_id=port['id']) [ 524.110647] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] nova.exception.PortBindingFailed: Binding failed for port e0651f92-1ada-45ee-b225-f8bf6b977110, please check neutron logs for more information. [ 524.110647] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] [ 524.110647] env[61867]: INFO nova.compute.manager [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Terminating instance [ 524.114146] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Acquiring lock "refresh_cache-dad6d190-9e74-452b-bb76-6d62f283218e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.114261] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Acquired lock "refresh_cache-dad6d190-9e74-452b-bb76-6d62f283218e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.114446] env[61867]: DEBUG nova.network.neutron [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 524.229633] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Acquiring lock "76d3a228-e314-42b1-9eec-43a32fbd8035" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.229862] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Lock "76d3a228-e314-42b1-9eec-43a32fbd8035" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.239526] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3be9d03-36fd-424d-93c2-90804950abcd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.249790] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15349070-892a-4eb0-8497-214c7d93a686 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.287267] env[61867]: DEBUG nova.policy [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '08ea6968be5a4a919da203f245dfbb25', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '222fa3f642ae4c63b95b130eca00d7e3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 524.288928] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04397b10-c462-4ebe-90f4-2597f9580623 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.297092] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c1eec0-154f-48f3-ac34-838905566bb1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.311825] env[61867]: DEBUG nova.compute.provider_tree [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 524.555885] env[61867]: DEBUG nova.compute.manager [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 524.628188] env[61867]: DEBUG nova.compute.manager [req-3d4b18bd-451e-4576-8a83-429edb60735d req-fc09303a-5873-4e6e-bb6a-a7cf75c1635f service nova] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Received event network-changed-e0651f92-1ada-45ee-b225-f8bf6b977110 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 524.628188] env[61867]: DEBUG nova.compute.manager [req-3d4b18bd-451e-4576-8a83-429edb60735d req-fc09303a-5873-4e6e-bb6a-a7cf75c1635f service nova] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Refreshing instance network info cache due to event network-changed-e0651f92-1ada-45ee-b225-f8bf6b977110. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 524.628188] env[61867]: DEBUG oslo_concurrency.lockutils [req-3d4b18bd-451e-4576-8a83-429edb60735d req-fc09303a-5873-4e6e-bb6a-a7cf75c1635f service nova] Acquiring lock "refresh_cache-dad6d190-9e74-452b-bb76-6d62f283218e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.658646] env[61867]: DEBUG nova.network.neutron [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.732451] env[61867]: DEBUG nova.compute.manager [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 524.816286] env[61867]: DEBUG nova.scheduler.client.report [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 524.975025] env[61867]: DEBUG nova.network.neutron [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.191973] env[61867]: DEBUG nova.compute.manager [req-ea122957-621e-4e91-a517-510ecb84576b req-82d3c1c4-733f-4983-81ef-fe51e1eac831 service nova] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Received event network-vif-deleted-f1202e09-9714-4663-884f-7f8053ddc16a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 525.267629] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.277042] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "a746fcff-646e-4f9b-b595-7379c6f97efd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.277348] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "a746fcff-646e-4f9b-b595-7379c6f97efd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.326568] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.788s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.326568] env[61867]: ERROR nova.compute.manager [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 982825fe-91fc-472a-b53a-30d50370f33a, please check neutron logs for more information. [ 525.326568] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Traceback (most recent call last): [ 525.326568] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 525.326568] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] self.driver.spawn(context, instance, image_meta, [ 525.326568] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 525.326568] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 525.326568] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 525.326568] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] vm_ref = self.build_virtual_machine(instance, [ 525.326898] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 525.326898] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] vif_infos = vmwarevif.get_vif_info(self._session, [ 525.326898] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 525.326898] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] for vif in network_info: [ 525.326898] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 525.326898] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] return self._sync_wrapper(fn, *args, **kwargs) [ 525.326898] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 525.326898] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] self.wait() [ 525.326898] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 525.326898] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] self[:] = self._gt.wait() [ 525.326898] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 525.326898] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] return self._exit_event.wait() [ 525.326898] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 525.327245] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] result = hub.switch() [ 525.327245] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 525.327245] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] return self.greenlet.switch() [ 525.327245] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 525.327245] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] result = function(*args, **kwargs) [ 525.327245] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 525.327245] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] return func(*args, **kwargs) [ 525.327245] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 525.327245] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] raise e [ 525.327245] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 525.327245] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] nwinfo = self.network_api.allocate_for_instance( [ 525.327245] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 525.327245] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] created_port_ids = self._update_ports_for_instance( [ 525.327697] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 525.327697] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] with excutils.save_and_reraise_exception(): [ 525.327697] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 525.327697] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] self.force_reraise() [ 525.327697] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 525.327697] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] raise self.value [ 525.327697] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 525.327697] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] updated_port = self._update_port( [ 525.327697] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 525.327697] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] _ensure_no_port_binding_failure(port) [ 525.327697] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 525.327697] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] raise exception.PortBindingFailed(port_id=port['id']) [ 525.328055] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] nova.exception.PortBindingFailed: Binding failed for port 982825fe-91fc-472a-b53a-30d50370f33a, please check neutron logs for more information. [ 525.328055] env[61867]: ERROR nova.compute.manager [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] [ 525.328055] env[61867]: DEBUG nova.compute.utils [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Binding failed for port 982825fe-91fc-472a-b53a-30d50370f33a, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 525.328460] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 5.318s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.328626] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.328783] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61867) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 525.329121] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 2.876s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.331925] env[61867]: DEBUG nova.compute.manager [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Build of instance ef0db253-4aba-44ee-9986-b015f383dfc5 was re-scheduled: Binding failed for port 982825fe-91fc-472a-b53a-30d50370f33a, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 525.334179] env[61867]: DEBUG nova.compute.manager [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 525.334179] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Acquiring lock "refresh_cache-ef0db253-4aba-44ee-9986-b015f383dfc5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.334179] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Acquired lock "refresh_cache-ef0db253-4aba-44ee-9986-b015f383dfc5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.334179] env[61867]: DEBUG nova.network.neutron [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 525.335526] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421422a0-1e26-48da-b78e-df759cbce832 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.346878] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d79afb6b-e066-41e3-a044-441d96198fde {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.370455] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9f353c-25f7-4a1b-ba61-d4f8db2fd040 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.381433] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd8bd9f-54a6-49d3-9649-cb5e777865f1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.417920] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181507MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61867) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 525.418025] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.484301] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Releasing lock "refresh_cache-dad6d190-9e74-452b-bb76-6d62f283218e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.484301] env[61867]: DEBUG nova.compute.manager [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 525.484301] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 525.484301] env[61867]: DEBUG oslo_concurrency.lockutils [req-3d4b18bd-451e-4576-8a83-429edb60735d req-fc09303a-5873-4e6e-bb6a-a7cf75c1635f service nova] Acquired lock "refresh_cache-dad6d190-9e74-452b-bb76-6d62f283218e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.484301] env[61867]: DEBUG nova.network.neutron [req-3d4b18bd-451e-4576-8a83-429edb60735d req-fc09303a-5873-4e6e-bb6a-a7cf75c1635f service nova] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Refreshing network info cache for port e0651f92-1ada-45ee-b225-f8bf6b977110 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 525.484522] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-027a68c0-927c-4388-aaf3-ab526e59caf8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.494875] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4493822-235f-431a-8f6e-f0d754143079 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.526803] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dad6d190-9e74-452b-bb76-6d62f283218e could not be found. [ 525.526803] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 525.526803] env[61867]: INFO nova.compute.manager [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 525.526803] env[61867]: DEBUG oslo.service.loopingcall [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 525.526803] env[61867]: DEBUG nova.compute.manager [-] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 525.526803] env[61867]: DEBUG nova.network.neutron [-] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 525.570632] env[61867]: DEBUG nova.compute.manager [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 525.598204] env[61867]: DEBUG nova.virt.hardware [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 525.598726] env[61867]: DEBUG nova.virt.hardware [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 525.598726] env[61867]: DEBUG nova.virt.hardware [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 525.598816] env[61867]: DEBUG nova.virt.hardware [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 525.598880] env[61867]: DEBUG nova.virt.hardware [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 525.599065] env[61867]: DEBUG nova.virt.hardware [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 525.599294] env[61867]: DEBUG nova.virt.hardware [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 525.599448] env[61867]: DEBUG nova.virt.hardware [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 525.599606] env[61867]: DEBUG nova.virt.hardware [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 525.599921] env[61867]: DEBUG nova.virt.hardware [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 525.600176] env[61867]: DEBUG nova.virt.hardware [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 525.601246] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48249b48-4f56-42fb-a51c-3a8d8b37d554 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.610096] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749d816c-cc08-4af5-b39e-b7a602a32121 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.691398] env[61867]: DEBUG nova.network.neutron [-] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 525.780101] env[61867]: DEBUG nova.compute.manager [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 525.876178] env[61867]: DEBUG nova.network.neutron [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 525.969930] env[61867]: DEBUG nova.network.neutron [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Successfully created port: 5bdbe391-5c75-46f4-8747-a36c9b236366 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 526.007640] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9ea1af-6db1-4216-81f2-9ed33b2270ee {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.016692] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d0439c2-e55c-4b55-b425-180f856d257a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.052994] env[61867]: DEBUG nova.network.neutron [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.059171] env[61867]: DEBUG nova.network.neutron [req-3d4b18bd-451e-4576-8a83-429edb60735d req-fc09303a-5873-4e6e-bb6a-a7cf75c1635f service nova] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 526.061866] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39536e6f-8f76-4041-803c-be96cb5c362e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.072068] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52005a28-f3d7-4c42-b608-efefe58f823b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.088226] env[61867]: DEBUG nova.compute.provider_tree [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 526.195313] env[61867]: DEBUG nova.network.neutron [-] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.307506] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.496859] env[61867]: DEBUG nova.network.neutron [req-3d4b18bd-451e-4576-8a83-429edb60735d req-fc09303a-5873-4e6e-bb6a-a7cf75c1635f service nova] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.560370] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Releasing lock "refresh_cache-ef0db253-4aba-44ee-9986-b015f383dfc5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 526.560370] env[61867]: DEBUG nova.compute.manager [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 526.560370] env[61867]: DEBUG nova.compute.manager [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 526.560533] env[61867]: DEBUG nova.network.neutron [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 526.591480] env[61867]: DEBUG nova.scheduler.client.report [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 526.615819] env[61867]: DEBUG nova.network.neutron [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 526.623921] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Acquiring lock "c3f0399f-543c-4ab7-8854-b5f7f012fdd2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.625708] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Lock "c3f0399f-543c-4ab7-8854-b5f7f012fdd2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.698153] env[61867]: INFO nova.compute.manager [-] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Took 1.17 seconds to deallocate network for instance. [ 526.702090] env[61867]: DEBUG nova.compute.claims [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 526.702308] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.999958] env[61867]: DEBUG oslo_concurrency.lockutils [req-3d4b18bd-451e-4576-8a83-429edb60735d req-fc09303a-5873-4e6e-bb6a-a7cf75c1635f service nova] Releasing lock "refresh_cache-dad6d190-9e74-452b-bb76-6d62f283218e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.099020] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.767s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.099020] env[61867]: ERROR nova.compute.manager [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f1202e09-9714-4663-884f-7f8053ddc16a, please check neutron logs for more information. [ 527.099020] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Traceback (most recent call last): [ 527.099020] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 527.099020] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] self.driver.spawn(context, instance, image_meta, [ 527.099020] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 527.099020] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] self._vmops.spawn(context, instance, image_meta, injected_files, [ 527.099020] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 527.099020] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] vm_ref = self.build_virtual_machine(instance, [ 527.099336] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 527.099336] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] vif_infos = vmwarevif.get_vif_info(self._session, [ 527.099336] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 527.099336] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] for vif in network_info: [ 527.099336] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 527.099336] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] return self._sync_wrapper(fn, *args, **kwargs) [ 527.099336] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 527.099336] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] self.wait() [ 527.099336] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 527.099336] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] self[:] = self._gt.wait() [ 527.099336] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 527.099336] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] return self._exit_event.wait() [ 527.099336] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 527.099925] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] result = hub.switch() [ 527.099925] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 527.099925] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] return self.greenlet.switch() [ 527.099925] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 527.099925] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] result = function(*args, **kwargs) [ 527.099925] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 527.099925] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] return func(*args, **kwargs) [ 527.099925] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 527.099925] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] raise e [ 527.099925] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.099925] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] nwinfo = self.network_api.allocate_for_instance( [ 527.099925] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 527.099925] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] created_port_ids = self._update_ports_for_instance( [ 527.100282] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 527.100282] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] with excutils.save_and_reraise_exception(): [ 527.100282] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.100282] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] self.force_reraise() [ 527.100282] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.100282] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] raise self.value [ 527.100282] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 527.100282] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] updated_port = self._update_port( [ 527.100282] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.100282] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] _ensure_no_port_binding_failure(port) [ 527.100282] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.100282] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] raise exception.PortBindingFailed(port_id=port['id']) [ 527.100576] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] nova.exception.PortBindingFailed: Binding failed for port f1202e09-9714-4663-884f-7f8053ddc16a, please check neutron logs for more information. [ 527.100576] env[61867]: ERROR nova.compute.manager [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] [ 527.100576] env[61867]: DEBUG nova.compute.utils [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Binding failed for port f1202e09-9714-4663-884f-7f8053ddc16a, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 527.101629] env[61867]: DEBUG nova.compute.manager [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Build of instance 510f3385-7dc0-4b73-b804-fa9e4f3ff995 was re-scheduled: Binding failed for port f1202e09-9714-4663-884f-7f8053ddc16a, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 527.101629] env[61867]: DEBUG nova.compute.manager [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 527.101818] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Acquiring lock "refresh_cache-510f3385-7dc0-4b73-b804-fa9e4f3ff995" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.101968] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Acquired lock "refresh_cache-510f3385-7dc0-4b73-b804-fa9e4f3ff995" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.102145] env[61867]: DEBUG nova.network.neutron [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 527.103227] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.579s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.107494] env[61867]: INFO nova.compute.claims [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 527.118746] env[61867]: DEBUG nova.network.neutron [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.129659] env[61867]: DEBUG nova.compute.manager [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 527.628045] env[61867]: INFO nova.compute.manager [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] Took 1.06 seconds to deallocate network for instance. [ 527.662650] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.688084] env[61867]: DEBUG nova.network.neutron [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 527.874717] env[61867]: DEBUG oslo_concurrency.lockutils [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Acquiring lock "629da2e8-2f09-42b7-b031-6d8f5a282e37" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.874944] env[61867]: DEBUG oslo_concurrency.lockutils [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Lock "629da2e8-2f09-42b7-b031-6d8f5a282e37" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.892679] env[61867]: DEBUG nova.network.neutron [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.986732] env[61867]: INFO nova.compute.manager [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Rebuilding instance [ 528.068868] env[61867]: DEBUG nova.compute.manager [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 528.069642] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-593e41ef-c205-4da9-87cc-b86066724ed9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.182669] env[61867]: ERROR nova.compute.manager [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f757b37f-f7bf-48b8-be0d-0ad5e3eadd6b, please check neutron logs for more information. [ 528.182669] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 528.182669] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 528.182669] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 528.182669] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 528.182669] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 528.182669] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 528.182669] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 528.182669] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.182669] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 528.182669] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.182669] env[61867]: ERROR nova.compute.manager raise self.value [ 528.182669] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 528.182669] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 528.182669] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.182669] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 528.183131] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.183131] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 528.183131] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f757b37f-f7bf-48b8-be0d-0ad5e3eadd6b, please check neutron logs for more information. [ 528.183131] env[61867]: ERROR nova.compute.manager [ 528.183131] env[61867]: Traceback (most recent call last): [ 528.183131] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 528.183131] env[61867]: listener.cb(fileno) [ 528.183131] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 528.183131] env[61867]: result = function(*args, **kwargs) [ 528.183131] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 528.183131] env[61867]: return func(*args, **kwargs) [ 528.183131] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 528.183131] env[61867]: raise e [ 528.183131] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 528.183131] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 528.183131] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 528.183131] env[61867]: created_port_ids = self._update_ports_for_instance( [ 528.183131] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 528.183131] env[61867]: with excutils.save_and_reraise_exception(): [ 528.183131] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.183131] env[61867]: self.force_reraise() [ 528.183131] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.183131] env[61867]: raise self.value [ 528.183131] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 528.183131] env[61867]: updated_port = self._update_port( [ 528.183131] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.183131] env[61867]: _ensure_no_port_binding_failure(port) [ 528.183131] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.183131] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 528.183799] env[61867]: nova.exception.PortBindingFailed: Binding failed for port f757b37f-f7bf-48b8-be0d-0ad5e3eadd6b, please check neutron logs for more information. [ 528.183799] env[61867]: Removing descriptor: 14 [ 528.183799] env[61867]: ERROR nova.compute.manager [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f757b37f-f7bf-48b8-be0d-0ad5e3eadd6b, please check neutron logs for more information. [ 528.183799] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Traceback (most recent call last): [ 528.183799] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 528.183799] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] yield resources [ 528.183799] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 528.183799] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] self.driver.spawn(context, instance, image_meta, [ 528.183799] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 528.183799] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 528.183799] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 528.183799] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] vm_ref = self.build_virtual_machine(instance, [ 528.184112] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 528.184112] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] vif_infos = vmwarevif.get_vif_info(self._session, [ 528.184112] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 528.184112] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] for vif in network_info: [ 528.184112] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 528.184112] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] return self._sync_wrapper(fn, *args, **kwargs) [ 528.184112] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 528.184112] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] self.wait() [ 528.184112] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 528.184112] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] self[:] = self._gt.wait() [ 528.184112] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 528.184112] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] return self._exit_event.wait() [ 528.184112] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 528.184448] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] result = hub.switch() [ 528.184448] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 528.184448] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] return self.greenlet.switch() [ 528.184448] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 528.184448] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] result = function(*args, **kwargs) [ 528.184448] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 528.184448] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] return func(*args, **kwargs) [ 528.184448] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 528.184448] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] raise e [ 528.184448] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 528.184448] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] nwinfo = self.network_api.allocate_for_instance( [ 528.184448] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 528.184448] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] created_port_ids = self._update_ports_for_instance( [ 528.184755] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 528.184755] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] with excutils.save_and_reraise_exception(): [ 528.184755] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.184755] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] self.force_reraise() [ 528.184755] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.184755] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] raise self.value [ 528.184755] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 528.184755] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] updated_port = self._update_port( [ 528.184755] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.184755] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] _ensure_no_port_binding_failure(port) [ 528.184755] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.184755] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] raise exception.PortBindingFailed(port_id=port['id']) [ 528.185270] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] nova.exception.PortBindingFailed: Binding failed for port f757b37f-f7bf-48b8-be0d-0ad5e3eadd6b, please check neutron logs for more information. [ 528.185270] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] [ 528.185270] env[61867]: INFO nova.compute.manager [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Terminating instance [ 528.191357] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Acquiring lock "refresh_cache-b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.191520] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Acquired lock "refresh_cache-b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.191686] env[61867]: DEBUG nova.network.neutron [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 528.370343] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4875860-0345-4d1e-ac62-3920e3763579 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.378283] env[61867]: DEBUG nova.compute.manager [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 528.385765] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88b0aa21-1d1b-40f8-8beb-6eac1888ee7a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.437628] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Releasing lock "refresh_cache-510f3385-7dc0-4b73-b804-fa9e4f3ff995" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.437908] env[61867]: DEBUG nova.compute.manager [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 528.438116] env[61867]: DEBUG nova.compute.manager [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 528.438263] env[61867]: DEBUG nova.network.neutron [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 528.441549] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0594eee3-fd1e-4f34-8784-6055ea544042 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.451833] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c4f910e-de69-4438-b8ba-9987cb4fee65 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.471463] env[61867]: DEBUG nova.compute.provider_tree [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 528.500347] env[61867]: DEBUG nova.network.neutron [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 528.587778] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 528.588074] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9a2e0fa9-0e2d-480e-b43a-76d529e62db8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.598079] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 528.598079] env[61867]: value = "task-1276298" [ 528.598079] env[61867]: _type = "Task" [ 528.598079] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.606958] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276298, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.674891] env[61867]: INFO nova.scheduler.client.report [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Deleted allocations for instance ef0db253-4aba-44ee-9986-b015f383dfc5 [ 528.733471] env[61867]: DEBUG nova.network.neutron [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 528.913257] env[61867]: DEBUG oslo_concurrency.lockutils [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.914148] env[61867]: DEBUG nova.network.neutron [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.977373] env[61867]: DEBUG nova.scheduler.client.report [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 529.003163] env[61867]: DEBUG nova.network.neutron [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.071767] env[61867]: DEBUG nova.compute.manager [req-f720e612-7f61-4828-af42-2f9c98096155 req-263394e6-b2ed-4e81-ae6a-318dc80ec32c service nova] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Received event network-vif-deleted-e0651f92-1ada-45ee-b225-f8bf6b977110 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 529.109462] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276298, 'name': PowerOffVM_Task, 'duration_secs': 0.125369} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.112258] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 529.112587] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 529.113367] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b17249-21a2-4cea-9410-ca7d1fe977c6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.121343] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 529.121960] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d29b635-19fc-407a-9956-3348b45ed702 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.152148] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 529.152807] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 529.152807] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Deleting the datastore file [datastore2] 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 529.152932] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-440fc3de-8ee5-4981-aae9-c4752f97d83e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.164488] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 529.164488] env[61867]: value = "task-1276300" [ 529.164488] env[61867]: _type = "Task" [ 529.164488] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.174759] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276300, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.188069] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff09bd76-8eb1-4903-a54f-0783b3ec41eb tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Lock "ef0db253-4aba-44ee-9986-b015f383dfc5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.218s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.188069] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "ef0db253-4aba-44ee-9986-b015f383dfc5" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 9.690s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.188069] env[61867]: INFO nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: ef0db253-4aba-44ee-9986-b015f383dfc5] During sync_power_state the instance has a pending task (spawning). Skip. [ 529.188265] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "ef0db253-4aba-44ee-9986-b015f383dfc5" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.418282] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Releasing lock "refresh_cache-b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 529.418911] env[61867]: DEBUG nova.compute.manager [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 529.418988] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 529.419275] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-be2f465b-89f3-4d24-bf31-41ca9dd51d93 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.429515] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e4ce08-db0c-4681-87a1-fdfb803aa218 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.458200] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7 could not be found. [ 529.460127] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 529.460127] env[61867]: INFO nova.compute.manager [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 529.460127] env[61867]: DEBUG oslo.service.loopingcall [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 529.460127] env[61867]: DEBUG nova.compute.manager [-] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 529.460127] env[61867]: DEBUG nova.network.neutron [-] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 529.483269] env[61867]: DEBUG nova.network.neutron [-] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 529.485077] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.382s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.487745] env[61867]: DEBUG nova.compute.manager [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 529.488342] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.221s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.489973] env[61867]: INFO nova.compute.claims [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 529.503841] env[61867]: ERROR nova.compute.manager [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5bdbe391-5c75-46f4-8747-a36c9b236366, please check neutron logs for more information. [ 529.503841] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 529.503841] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.503841] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 529.503841] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 529.503841] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 529.503841] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 529.503841] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 529.503841] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.503841] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 529.503841] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.503841] env[61867]: ERROR nova.compute.manager raise self.value [ 529.503841] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 529.503841] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 529.503841] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.503841] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 529.504330] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.504330] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 529.504330] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5bdbe391-5c75-46f4-8747-a36c9b236366, please check neutron logs for more information. [ 529.504330] env[61867]: ERROR nova.compute.manager [ 529.504330] env[61867]: Traceback (most recent call last): [ 529.504330] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 529.504330] env[61867]: listener.cb(fileno) [ 529.504330] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.504330] env[61867]: result = function(*args, **kwargs) [ 529.504330] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 529.504330] env[61867]: return func(*args, **kwargs) [ 529.504330] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.504330] env[61867]: raise e [ 529.504330] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.504330] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 529.504330] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 529.504330] env[61867]: created_port_ids = self._update_ports_for_instance( [ 529.504330] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 529.504330] env[61867]: with excutils.save_and_reraise_exception(): [ 529.504330] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.504330] env[61867]: self.force_reraise() [ 529.504330] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.504330] env[61867]: raise self.value [ 529.504330] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 529.504330] env[61867]: updated_port = self._update_port( [ 529.504330] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.504330] env[61867]: _ensure_no_port_binding_failure(port) [ 529.504330] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.504330] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 529.505131] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 5bdbe391-5c75-46f4-8747-a36c9b236366, please check neutron logs for more information. [ 529.505131] env[61867]: Removing descriptor: 17 [ 529.505131] env[61867]: ERROR nova.compute.manager [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5bdbe391-5c75-46f4-8747-a36c9b236366, please check neutron logs for more information. [ 529.505131] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Traceback (most recent call last): [ 529.505131] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 529.505131] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] yield resources [ 529.505131] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 529.505131] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] self.driver.spawn(context, instance, image_meta, [ 529.505131] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 529.505131] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 529.505131] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 529.505131] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] vm_ref = self.build_virtual_machine(instance, [ 529.505522] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 529.505522] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 529.505522] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 529.505522] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] for vif in network_info: [ 529.505522] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 529.505522] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] return self._sync_wrapper(fn, *args, **kwargs) [ 529.505522] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 529.505522] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] self.wait() [ 529.505522] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 529.505522] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] self[:] = self._gt.wait() [ 529.505522] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 529.505522] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] return self._exit_event.wait() [ 529.505522] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 529.505862] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] result = hub.switch() [ 529.505862] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 529.505862] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] return self.greenlet.switch() [ 529.505862] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.505862] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] result = function(*args, **kwargs) [ 529.505862] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 529.505862] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] return func(*args, **kwargs) [ 529.505862] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.505862] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] raise e [ 529.505862] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.505862] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] nwinfo = self.network_api.allocate_for_instance( [ 529.505862] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 529.505862] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] created_port_ids = self._update_ports_for_instance( [ 529.506319] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 529.506319] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] with excutils.save_and_reraise_exception(): [ 529.506319] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.506319] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] self.force_reraise() [ 529.506319] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.506319] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] raise self.value [ 529.506319] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 529.506319] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] updated_port = self._update_port( [ 529.506319] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.506319] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] _ensure_no_port_binding_failure(port) [ 529.506319] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.506319] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] raise exception.PortBindingFailed(port_id=port['id']) [ 529.506617] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] nova.exception.PortBindingFailed: Binding failed for port 5bdbe391-5c75-46f4-8747-a36c9b236366, please check neutron logs for more information. [ 529.506617] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] [ 529.506617] env[61867]: INFO nova.compute.manager [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Terminating instance [ 529.510872] env[61867]: INFO nova.compute.manager [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] Took 1.07 seconds to deallocate network for instance. [ 529.516413] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Acquiring lock "refresh_cache-433ce856-b04b-4808-8f05-201ad28e18d0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.517388] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Acquired lock "refresh_cache-433ce856-b04b-4808-8f05-201ad28e18d0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.517388] env[61867]: DEBUG nova.network.neutron [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 529.679626] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276300, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.099928} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.680095] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 529.680777] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 529.681026] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 529.742247] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Acquiring lock "10bd5056-d5b5-48d7-a5b5-bd07dd489618" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.742247] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Lock "10bd5056-d5b5-48d7-a5b5-bd07dd489618" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.965019] env[61867]: DEBUG nova.compute.manager [req-7af8a066-8ff5-49c2-9798-ba48a898f097 req-c8853d57-aab7-4081-a388-58f651b5ec51 service nova] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Received event network-changed-f757b37f-f7bf-48b8-be0d-0ad5e3eadd6b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 529.965248] env[61867]: DEBUG nova.compute.manager [req-7af8a066-8ff5-49c2-9798-ba48a898f097 req-c8853d57-aab7-4081-a388-58f651b5ec51 service nova] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Refreshing instance network info cache due to event network-changed-f757b37f-f7bf-48b8-be0d-0ad5e3eadd6b. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 529.965466] env[61867]: DEBUG oslo_concurrency.lockutils [req-7af8a066-8ff5-49c2-9798-ba48a898f097 req-c8853d57-aab7-4081-a388-58f651b5ec51 service nova] Acquiring lock "refresh_cache-b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.965611] env[61867]: DEBUG oslo_concurrency.lockutils [req-7af8a066-8ff5-49c2-9798-ba48a898f097 req-c8853d57-aab7-4081-a388-58f651b5ec51 service nova] Acquired lock "refresh_cache-b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.965839] env[61867]: DEBUG nova.network.neutron [req-7af8a066-8ff5-49c2-9798-ba48a898f097 req-c8853d57-aab7-4081-a388-58f651b5ec51 service nova] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Refreshing network info cache for port f757b37f-f7bf-48b8-be0d-0ad5e3eadd6b {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 529.988775] env[61867]: DEBUG nova.network.neutron [-] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.993590] env[61867]: DEBUG nova.compute.utils [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 530.003075] env[61867]: DEBUG nova.compute.manager [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 530.003275] env[61867]: DEBUG nova.network.neutron [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 530.084017] env[61867]: DEBUG nova.network.neutron [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 530.127298] env[61867]: DEBUG nova.policy [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd6f1a604e0149dda8648a7f0aa63c44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f877e6e0b8ae41d7b6cae781df60020f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 530.244085] env[61867]: DEBUG nova.compute.manager [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 530.332206] env[61867]: DEBUG nova.network.neutron [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.491081] env[61867]: INFO nova.compute.manager [-] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Took 1.03 seconds to deallocate network for instance. [ 530.498039] env[61867]: DEBUG nova.compute.claims [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 530.498039] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.498279] env[61867]: DEBUG nova.network.neutron [req-7af8a066-8ff5-49c2-9798-ba48a898f097 req-c8853d57-aab7-4081-a388-58f651b5ec51 service nova] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 530.505040] env[61867]: DEBUG nova.compute.manager [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 530.550752] env[61867]: INFO nova.scheduler.client.report [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Deleted allocations for instance 510f3385-7dc0-4b73-b804-fa9e4f3ff995 [ 530.706452] env[61867]: DEBUG nova.network.neutron [req-7af8a066-8ff5-49c2-9798-ba48a898f097 req-c8853d57-aab7-4081-a388-58f651b5ec51 service nova] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.727051] env[61867]: DEBUG nova.virt.hardware [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 530.727997] env[61867]: DEBUG nova.virt.hardware [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 530.727997] env[61867]: DEBUG nova.virt.hardware [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 530.728307] env[61867]: DEBUG nova.virt.hardware [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 530.728572] env[61867]: DEBUG nova.virt.hardware [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 530.729121] env[61867]: DEBUG nova.virt.hardware [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 530.729121] env[61867]: DEBUG nova.virt.hardware [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 530.729282] env[61867]: DEBUG nova.virt.hardware [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 530.729494] env[61867]: DEBUG nova.virt.hardware [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 530.729698] env[61867]: DEBUG nova.virt.hardware [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 530.729894] env[61867]: DEBUG nova.virt.hardware [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 530.731044] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f0656e-b1c5-4d30-ab6b-23dd82c9c168 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.735473] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52d4be8-7276-4cb6-810e-ee960d0069e5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.747079] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0054b62-6665-45ea-840e-4bde27e2163b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.754408] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78fc2334-b479-40fe-9d9a-a3e7ecad3d74 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.771745] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Instance VIF info [] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 530.778153] env[61867]: DEBUG oslo.service.loopingcall [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 530.779521] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.780433] env[61867]: DEBUG nova.network.neutron [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Successfully created port: b3b1de3c-67f3-4a32-a601-90f3081ce672 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 530.807844] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 530.811019] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4c2c7751-16ad-459f-9abe-b8f6840ffeee {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.825062] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8ba603-c051-43b5-8ec6-9711e20ea176 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.834672] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Releasing lock "refresh_cache-433ce856-b04b-4808-8f05-201ad28e18d0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 530.835803] env[61867]: DEBUG nova.compute.manager [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 530.837119] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 530.837544] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0512960c-a367-49ab-aff8-7db8e89af52a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.842247] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-090b106d-f500-4527-931c-23d0a29d2104 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.846504] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 530.846504] env[61867]: value = "task-1276301" [ 530.846504] env[61867]: _type = "Task" [ 530.846504] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.862350] env[61867]: DEBUG nova.compute.provider_tree [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 530.868340] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5891786a-0cde-48be-87a6-62a19aaea7a2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.888681] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276301, 'name': CreateVM_Task} progress is 15%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.896743] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 433ce856-b04b-4808-8f05-201ad28e18d0 could not be found. [ 530.896910] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 530.897035] env[61867]: INFO nova.compute.manager [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Took 0.06 seconds to destroy the instance on the hypervisor. [ 530.897256] env[61867]: DEBUG oslo.service.loopingcall [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 530.897475] env[61867]: DEBUG nova.compute.manager [-] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 530.897572] env[61867]: DEBUG nova.network.neutron [-] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 530.930106] env[61867]: DEBUG nova.network.neutron [-] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 531.067805] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e73ea9c-42b0-4863-aa8a-4ed36a9e6e35 tempest-ServersAdminNegativeTestJSON-871909674 tempest-ServersAdminNegativeTestJSON-871909674-project-member] Lock "510f3385-7dc0-4b73-b804-fa9e4f3ff995" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.165s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.068076] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "510f3385-7dc0-4b73-b804-fa9e4f3ff995" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 11.570s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.068270] env[61867]: INFO nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 510f3385-7dc0-4b73-b804-fa9e4f3ff995] During sync_power_state the instance has a pending task (spawning). Skip. [ 531.068445] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "510f3385-7dc0-4b73-b804-fa9e4f3ff995" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.212631] env[61867]: DEBUG oslo_concurrency.lockutils [req-7af8a066-8ff5-49c2-9798-ba48a898f097 req-c8853d57-aab7-4081-a388-58f651b5ec51 service nova] Releasing lock "refresh_cache-b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.360603] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276301, 'name': CreateVM_Task, 'duration_secs': 0.323701} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.360769] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 531.361221] env[61867]: DEBUG oslo_concurrency.lockutils [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 531.361371] env[61867]: DEBUG oslo_concurrency.lockutils [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 531.361689] env[61867]: DEBUG oslo_concurrency.lockutils [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 531.361933] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5eabede9-ae71-47b7-b683-cbd550e889ad {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.368332] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 531.368332] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527de5be-83b1-5d1f-aa3c-ec081e72c524" [ 531.368332] env[61867]: _type = "Task" [ 531.368332] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.377921] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527de5be-83b1-5d1f-aa3c-ec081e72c524, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.382244] env[61867]: DEBUG nova.scheduler.client.report [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 531.433461] env[61867]: DEBUG nova.network.neutron [-] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.514769] env[61867]: DEBUG nova.compute.manager [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 531.549026] env[61867]: DEBUG nova.virt.hardware [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 531.549807] env[61867]: DEBUG nova.virt.hardware [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 531.553166] env[61867]: DEBUG nova.virt.hardware [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 531.553166] env[61867]: DEBUG nova.virt.hardware [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 531.553166] env[61867]: DEBUG nova.virt.hardware [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 531.553166] env[61867]: DEBUG nova.virt.hardware [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 531.553166] env[61867]: DEBUG nova.virt.hardware [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 531.557137] env[61867]: DEBUG nova.virt.hardware [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 531.557137] env[61867]: DEBUG nova.virt.hardware [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 531.557137] env[61867]: DEBUG nova.virt.hardware [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 531.557137] env[61867]: DEBUG nova.virt.hardware [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 531.557137] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d4e85c-a70d-4df9-a8a9-7a46bf2b6bbd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.563174] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c82ae33-e3ed-42e9-89c8-6917dbf7e1a8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.879837] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527de5be-83b1-5d1f-aa3c-ec081e72c524, 'name': SearchDatastore_Task, 'duration_secs': 0.016705} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.880173] env[61867]: DEBUG oslo_concurrency.lockutils [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.880390] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 531.880623] env[61867]: DEBUG oslo_concurrency.lockutils [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 531.880765] env[61867]: DEBUG oslo_concurrency.lockutils [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 531.880934] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 531.881200] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dad22ab3-6d15-44cf-9029-58d0715a3678 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.890186] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 531.890701] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 531.891630] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.892113] env[61867]: DEBUG nova.compute.manager [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 531.894682] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-043561bf-674d-408c-a43b-d1dca362cd5e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.897389] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 6.479s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.902973] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 531.902973] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5228a008-47b0-d477-7408-ceae7a16c9ce" [ 531.902973] env[61867]: _type = "Task" [ 531.902973] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.912639] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5228a008-47b0-d477-7408-ceae7a16c9ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.937347] env[61867]: INFO nova.compute.manager [-] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Took 1.04 seconds to deallocate network for instance. [ 531.939968] env[61867]: DEBUG nova.compute.claims [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 531.940173] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.108988] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Acquiring lock "11292ffa-0db5-4d70-a3c1-31b81bf4182e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.109433] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Lock "11292ffa-0db5-4d70-a3c1-31b81bf4182e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.399273] env[61867]: DEBUG nova.compute.utils [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 532.405022] env[61867]: DEBUG nova.compute.manager [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Not allocating networking since 'none' was specified. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 532.426452] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5228a008-47b0-d477-7408-ceae7a16c9ce, 'name': SearchDatastore_Task, 'duration_secs': 0.008787} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.427281] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b119cf6d-4496-4abd-8697-cb469b2f3d73 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.433567] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 532.433567] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52078ef9-6dcc-e492-8b97-edd09c7d2b79" [ 532.433567] env[61867]: _type = "Task" [ 532.433567] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.444610] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52078ef9-6dcc-e492-8b97-edd09c7d2b79, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.474841] env[61867]: ERROR nova.compute.manager [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b3b1de3c-67f3-4a32-a601-90f3081ce672, please check neutron logs for more information. [ 532.474841] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 532.474841] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.474841] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 532.474841] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 532.474841] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 532.474841] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 532.474841] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 532.474841] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.474841] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 532.474841] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.474841] env[61867]: ERROR nova.compute.manager raise self.value [ 532.474841] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 532.474841] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 532.474841] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.474841] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 532.475318] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.475318] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 532.475318] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b3b1de3c-67f3-4a32-a601-90f3081ce672, please check neutron logs for more information. [ 532.475318] env[61867]: ERROR nova.compute.manager [ 532.475318] env[61867]: Traceback (most recent call last): [ 532.475457] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 532.475457] env[61867]: listener.cb(fileno) [ 532.475457] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.475457] env[61867]: result = function(*args, **kwargs) [ 532.475457] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 532.475457] env[61867]: return func(*args, **kwargs) [ 532.475457] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.475457] env[61867]: raise e [ 532.475457] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.475457] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 532.475457] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 532.475457] env[61867]: created_port_ids = self._update_ports_for_instance( [ 532.475457] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 532.475457] env[61867]: with excutils.save_and_reraise_exception(): [ 532.475457] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.475457] env[61867]: self.force_reraise() [ 532.475457] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.475457] env[61867]: raise self.value [ 532.475457] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 532.475457] env[61867]: updated_port = self._update_port( [ 532.475457] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.475457] env[61867]: _ensure_no_port_binding_failure(port) [ 532.475457] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.475457] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 532.475457] env[61867]: nova.exception.PortBindingFailed: Binding failed for port b3b1de3c-67f3-4a32-a601-90f3081ce672, please check neutron logs for more information. [ 532.475457] env[61867]: Removing descriptor: 17 [ 532.477628] env[61867]: ERROR nova.compute.manager [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b3b1de3c-67f3-4a32-a601-90f3081ce672, please check neutron logs for more information. [ 532.477628] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Traceback (most recent call last): [ 532.477628] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 532.477628] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] yield resources [ 532.477628] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 532.477628] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] self.driver.spawn(context, instance, image_meta, [ 532.477628] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 532.477628] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 532.477628] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 532.477628] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] vm_ref = self.build_virtual_machine(instance, [ 532.477628] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 532.478031] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] vif_infos = vmwarevif.get_vif_info(self._session, [ 532.478031] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 532.478031] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] for vif in network_info: [ 532.478031] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 532.478031] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] return self._sync_wrapper(fn, *args, **kwargs) [ 532.478031] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 532.478031] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] self.wait() [ 532.478031] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 532.478031] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] self[:] = self._gt.wait() [ 532.478031] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 532.478031] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] return self._exit_event.wait() [ 532.478031] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 532.478031] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] result = hub.switch() [ 532.478414] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 532.478414] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] return self.greenlet.switch() [ 532.478414] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.478414] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] result = function(*args, **kwargs) [ 532.478414] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 532.478414] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] return func(*args, **kwargs) [ 532.478414] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.478414] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] raise e [ 532.478414] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.478414] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] nwinfo = self.network_api.allocate_for_instance( [ 532.478414] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 532.478414] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] created_port_ids = self._update_ports_for_instance( [ 532.478414] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 532.478734] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] with excutils.save_and_reraise_exception(): [ 532.478734] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.478734] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] self.force_reraise() [ 532.478734] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.478734] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] raise self.value [ 532.478734] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 532.478734] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] updated_port = self._update_port( [ 532.478734] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.478734] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] _ensure_no_port_binding_failure(port) [ 532.478734] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.478734] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] raise exception.PortBindingFailed(port_id=port['id']) [ 532.478734] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] nova.exception.PortBindingFailed: Binding failed for port b3b1de3c-67f3-4a32-a601-90f3081ce672, please check neutron logs for more information. [ 532.478734] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] [ 532.479867] env[61867]: INFO nova.compute.manager [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Terminating instance [ 532.482335] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Acquiring lock "refresh_cache-ac5df149-83db-481a-a240-e5cf08fdbff9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.482410] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Acquired lock "refresh_cache-ac5df149-83db-481a-a240-e5cf08fdbff9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.482525] env[61867]: DEBUG nova.network.neutron [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 532.612138] env[61867]: DEBUG nova.compute.manager [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 532.908731] env[61867]: DEBUG nova.compute.manager [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 532.947626] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 532.948350] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance dad6d190-9e74-452b-bb76-6d62f283218e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 532.948350] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 433ce856-b04b-4808-8f05-201ad28e18d0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 532.948350] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 532.948350] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance ac5df149-83db-481a-a240-e5cf08fdbff9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 532.948551] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 76d3a228-e314-42b1-9eec-43a32fbd8035 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 532.954580] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52078ef9-6dcc-e492-8b97-edd09c7d2b79, 'name': SearchDatastore_Task, 'duration_secs': 0.009104} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.955388] env[61867]: DEBUG oslo_concurrency.lockutils [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 532.955857] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa/318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 532.956119] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e77d03af-39c6-4ffd-880c-2b3fc889ce33 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.966128] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 532.966128] env[61867]: value = "task-1276302" [ 532.966128] env[61867]: _type = "Task" [ 532.966128] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.982391] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276302, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.007637] env[61867]: DEBUG nova.network.neutron [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 533.096172] env[61867]: DEBUG nova.network.neutron [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.149604] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.457180] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance a746fcff-646e-4f9b-b595-7379c6f97efd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 533.478039] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276302, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454356} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.481660] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa/318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 533.481660] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 533.481660] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-365f2fb1-80c6-462a-af8f-1e944780174e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.488501] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 533.488501] env[61867]: value = "task-1276303" [ 533.488501] env[61867]: _type = "Task" [ 533.488501] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.500645] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276303, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.598945] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Releasing lock "refresh_cache-ac5df149-83db-481a-a240-e5cf08fdbff9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.598945] env[61867]: DEBUG nova.compute.manager [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 533.598945] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 533.599238] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e9ecc85d-87b5-41e5-8b76-2926e857bbb3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.620339] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dce306d-ba59-482d-a82f-18e89d7ae79c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.651048] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ac5df149-83db-481a-a240-e5cf08fdbff9 could not be found. [ 533.651283] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 533.651528] env[61867]: INFO nova.compute.manager [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Took 0.05 seconds to destroy the instance on the hypervisor. [ 533.651815] env[61867]: DEBUG oslo.service.loopingcall [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 533.652112] env[61867]: DEBUG nova.compute.manager [-] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 533.652207] env[61867]: DEBUG nova.network.neutron [-] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 533.669914] env[61867]: DEBUG nova.network.neutron [-] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 533.922635] env[61867]: DEBUG nova.compute.manager [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 533.960282] env[61867]: DEBUG nova.virt.hardware [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 533.960743] env[61867]: DEBUG nova.virt.hardware [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 533.961479] env[61867]: DEBUG nova.virt.hardware [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 533.961479] env[61867]: DEBUG nova.virt.hardware [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 533.961612] env[61867]: DEBUG nova.virt.hardware [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 533.962294] env[61867]: DEBUG nova.virt.hardware [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 533.962294] env[61867]: DEBUG nova.virt.hardware [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 533.962294] env[61867]: DEBUG nova.virt.hardware [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 533.962294] env[61867]: DEBUG nova.virt.hardware [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 533.962483] env[61867]: DEBUG nova.virt.hardware [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 533.962597] env[61867]: DEBUG nova.virt.hardware [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 533.963336] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance c3f0399f-543c-4ab7-8854-b5f7f012fdd2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 533.965656] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aae7f58-92ae-4ac9-9fcb-c49a4a1b3544 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.977623] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad86de57-e023-40c0-a4de-34372b668d0c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.998596] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Instance VIF info [] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 534.007183] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Creating folder: Project (1b6d2359353f4f77af9bd5bca11f37ef). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 534.012010] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cc3047f9-73e0-4c32-9777-45e4eaf472b4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.018190] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276303, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062085} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.018190] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 534.018963] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f931e78-9c33-4681-ad8d-530c8e59f18f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.022679] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Created folder: Project (1b6d2359353f4f77af9bd5bca11f37ef) in parent group-v274258. [ 534.023610] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Creating folder: Instances. Parent ref: group-v274267. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 534.023610] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6ce11729-ced8-4107-b4f0-0ced90603292 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.041932] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa/318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 534.043988] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0106bcd8-74e8-4a77-add6-c050ae28f3a1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.060242] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Created folder: Instances in parent group-v274267. [ 534.060458] env[61867]: DEBUG oslo.service.loopingcall [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 534.060924] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 534.061401] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4135f9f6-2e73-4844-b11d-e331693c75a5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.075441] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 534.075441] env[61867]: value = "task-1276306" [ 534.075441] env[61867]: _type = "Task" [ 534.075441] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.080442] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 534.080442] env[61867]: value = "task-1276307" [ 534.080442] env[61867]: _type = "Task" [ 534.080442] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.087868] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276306, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.093399] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276307, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.173750] env[61867]: DEBUG nova.network.neutron [-] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.471153] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 629da2e8-2f09-42b7-b031-6d8f5a282e37 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 534.589487] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276306, 'name': ReconfigVM_Task, 'duration_secs': 0.336858} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.592626] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa/318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 534.593192] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276307, 'name': CreateVM_Task, 'duration_secs': 0.285861} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.593371] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ce6287b7-f9a6-4ba4-8ea6-ccb4531c106b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.595090] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 534.595892] env[61867]: DEBUG oslo_vmware.service [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09a362b4-95e6-49ac-91a6-abf669df148b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.602421] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.602580] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.602930] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 534.603158] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f11fb164-b22a-43a8-a751-1512c5059c89 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.605802] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 534.605802] env[61867]: value = "task-1276308" [ 534.605802] env[61867]: _type = "Task" [ 534.605802] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.611200] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Waiting for the task: (returnval){ [ 534.611200] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5293fb31-144e-ed94-9baf-d497e2eef5c1" [ 534.611200] env[61867]: _type = "Task" [ 534.611200] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.617800] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276308, 'name': Rename_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.622409] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5293fb31-144e-ed94-9baf-d497e2eef5c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.680994] env[61867]: INFO nova.compute.manager [-] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Took 1.03 seconds to deallocate network for instance. [ 534.684060] env[61867]: DEBUG nova.compute.claims [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 534.684269] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.888034] env[61867]: DEBUG oslo_concurrency.lockutils [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Acquiring lock "f911465c-277a-41ee-a6b5-0e3b85185b56" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.888034] env[61867]: DEBUG oslo_concurrency.lockutils [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Lock "f911465c-277a-41ee-a6b5-0e3b85185b56" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.974264] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 10bd5056-d5b5-48d7-a5b5-bd07dd489618 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 535.122449] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276308, 'name': Rename_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.128687] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.130390] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 535.130390] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.130390] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.130390] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 535.130390] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff1def68-6faf-49f7-9dd8-7ae3bebdbf26 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.143280] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 535.143280] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 535.143280] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32df3a38-4cff-463a-ab48-93b73900a6dc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.152017] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da34009c-2445-4150-a371-dd0eb8fb3d49 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.156844] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Waiting for the task: (returnval){ [ 535.156844] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529a09ae-5086-4835-e172-f769318fcc0a" [ 535.156844] env[61867]: _type = "Task" [ 535.156844] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.164841] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529a09ae-5086-4835-e172-f769318fcc0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.365534] env[61867]: DEBUG nova.compute.manager [req-6d9af304-fbad-4161-aa2d-3cfdbb4165df req-af4335a1-7f9a-48cf-baaf-59f16061cd98 service nova] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Received event network-vif-deleted-f757b37f-f7bf-48b8-be0d-0ad5e3eadd6b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 535.365534] env[61867]: DEBUG nova.compute.manager [req-6d9af304-fbad-4161-aa2d-3cfdbb4165df req-af4335a1-7f9a-48cf-baaf-59f16061cd98 service nova] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Received event network-changed-5bdbe391-5c75-46f4-8747-a36c9b236366 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 535.365763] env[61867]: DEBUG nova.compute.manager [req-6d9af304-fbad-4161-aa2d-3cfdbb4165df req-af4335a1-7f9a-48cf-baaf-59f16061cd98 service nova] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Refreshing instance network info cache due to event network-changed-5bdbe391-5c75-46f4-8747-a36c9b236366. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 535.366066] env[61867]: DEBUG oslo_concurrency.lockutils [req-6d9af304-fbad-4161-aa2d-3cfdbb4165df req-af4335a1-7f9a-48cf-baaf-59f16061cd98 service nova] Acquiring lock "refresh_cache-433ce856-b04b-4808-8f05-201ad28e18d0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.366278] env[61867]: DEBUG oslo_concurrency.lockutils [req-6d9af304-fbad-4161-aa2d-3cfdbb4165df req-af4335a1-7f9a-48cf-baaf-59f16061cd98 service nova] Acquired lock "refresh_cache-433ce856-b04b-4808-8f05-201ad28e18d0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.366517] env[61867]: DEBUG nova.network.neutron [req-6d9af304-fbad-4161-aa2d-3cfdbb4165df req-af4335a1-7f9a-48cf-baaf-59f16061cd98 service nova] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Refreshing network info cache for port 5bdbe391-5c75-46f4-8747-a36c9b236366 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 535.476987] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 11292ffa-0db5-4d70-a3c1-31b81bf4182e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 535.477266] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 535.477415] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 535.623915] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276308, 'name': Rename_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.673716] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Preparing fetch location {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 535.673974] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Creating directory with path [datastore1] vmware_temp/26e44410-c72a-474e-b6fc-16d791982618/4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 535.674222] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fffc0beb-a296-48c9-bf48-ac9022bf4e76 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.699188] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Created directory with path [datastore1] vmware_temp/26e44410-c72a-474e-b6fc-16d791982618/4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 535.699188] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Fetch image to [datastore1] vmware_temp/26e44410-c72a-474e-b6fc-16d791982618/4ca02567-c128-482f-b204-6ac166ac3160/tmp-sparse.vmdk {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 535.699188] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Downloading image file data 4ca02567-c128-482f-b204-6ac166ac3160 to [datastore1] vmware_temp/26e44410-c72a-474e-b6fc-16d791982618/4ca02567-c128-482f-b204-6ac166ac3160/tmp-sparse.vmdk on the data store datastore1 {{(pid=61867) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 535.699188] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e9c28d-d5ac-4ee5-bdf8-557a117fdb55 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.702522] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5ffccd3-29b5-4a21-a4a9-36b8b85a91eb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.711850] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f60275-6ee1-4fba-b5d5-f5e5563f69b8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.714991] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88cd08df-cb9b-4e1e-962b-0fe33c8b6e90 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.751336] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a855ef-f887-4379-af2a-2b8b726182ca {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.759370] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9810dfa8-b0f4-464f-9612-bf694f210ab9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.794208] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e68b6a48-c676-4206-a2f7-404af227c0b0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.798940] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d8c550-725b-4679-a225-5908b27815cc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.812283] env[61867]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-0af1af8f-1075-4d43-ad13-e26dfa17b6a7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.814103] env[61867]: DEBUG nova.compute.provider_tree [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 535.887887] env[61867]: DEBUG nova.network.neutron [req-6d9af304-fbad-4161-aa2d-3cfdbb4165df req-af4335a1-7f9a-48cf-baaf-59f16061cd98 service nova] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 535.902109] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Downloading image file data 4ca02567-c128-482f-b204-6ac166ac3160 to the data store datastore1 {{(pid=61867) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 535.970858] env[61867]: DEBUG oslo_vmware.rw_handles [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/26e44410-c72a-474e-b6fc-16d791982618/4ca02567-c128-482f-b204-6ac166ac3160/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61867) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 536.038156] env[61867]: DEBUG nova.network.neutron [req-6d9af304-fbad-4161-aa2d-3cfdbb4165df req-af4335a1-7f9a-48cf-baaf-59f16061cd98 service nova] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.125780] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276308, 'name': Rename_Task, 'duration_secs': 1.124273} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.126114] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 536.126363] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b4d6af5e-cd66-4256-8575-80163cb7f7c7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.134435] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 536.134435] env[61867]: value = "task-1276309" [ 536.134435] env[61867]: _type = "Task" [ 536.134435] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.148993] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276309, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.316632] env[61867]: DEBUG nova.scheduler.client.report [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 536.542868] env[61867]: DEBUG oslo_concurrency.lockutils [req-6d9af304-fbad-4161-aa2d-3cfdbb4165df req-af4335a1-7f9a-48cf-baaf-59f16061cd98 service nova] Releasing lock "refresh_cache-433ce856-b04b-4808-8f05-201ad28e18d0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.542868] env[61867]: DEBUG nova.compute.manager [req-6d9af304-fbad-4161-aa2d-3cfdbb4165df req-af4335a1-7f9a-48cf-baaf-59f16061cd98 service nova] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Received event network-vif-deleted-5bdbe391-5c75-46f4-8747-a36c9b236366 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 536.648780] env[61867]: DEBUG oslo_vmware.api [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276309, 'name': PowerOnVM_Task, 'duration_secs': 0.457645} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.654089] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 536.654438] env[61867]: DEBUG nova.compute.manager [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 536.655188] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c49ddae0-1196-433e-b86c-a1d298c3069e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.694943] env[61867]: DEBUG oslo_vmware.rw_handles [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Completed reading data from the image iterator. {{(pid=61867) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 536.695142] env[61867]: DEBUG oslo_vmware.rw_handles [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Closing write handle for https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/26e44410-c72a-474e-b6fc-16d791982618/4ca02567-c128-482f-b204-6ac166ac3160/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61867) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 536.823301] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61867) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 536.823860] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.926s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 536.824194] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.517s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.827930] env[61867]: INFO nova.compute.claims [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 536.831263] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 536.831368] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Getting list of instances from cluster (obj){ [ 536.831368] env[61867]: value = "domain-c8" [ 536.831368] env[61867]: _type = "ClusterComputeResource" [ 536.831368] env[61867]: } {{(pid=61867) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 536.833967] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Downloaded image file data 4ca02567-c128-482f-b204-6ac166ac3160 to vmware_temp/26e44410-c72a-474e-b6fc-16d791982618/4ca02567-c128-482f-b204-6ac166ac3160/tmp-sparse.vmdk on the data store datastore1 {{(pid=61867) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 536.836866] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Caching image {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 536.836866] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Copying Virtual Disk [datastore1] vmware_temp/26e44410-c72a-474e-b6fc-16d791982618/4ca02567-c128-482f-b204-6ac166ac3160/tmp-sparse.vmdk to [datastore1] vmware_temp/26e44410-c72a-474e-b6fc-16d791982618/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 536.839839] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08f9278-6fc6-4a35-bf80-6f08bfd431bf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.844061] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-06f3b60b-4c74-4ff3-bc63-0a90aad58cc3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.856425] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Waiting for the task: (returnval){ [ 536.856425] env[61867]: value = "task-1276310" [ 536.856425] env[61867]: _type = "Task" [ 536.856425] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.862200] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Got total of 2 instances {{(pid=61867) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 536.875815] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276310, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.176325] env[61867]: DEBUG oslo_concurrency.lockutils [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.377127] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276310, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.409571] env[61867]: DEBUG oslo_concurrency.lockutils [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Acquiring lock "814dd6d2-43b1-4700-b585-c5d33b96931f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.409571] env[61867]: DEBUG oslo_concurrency.lockutils [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Lock "814dd6d2-43b1-4700-b585-c5d33b96931f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.442187] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Acquiring lock "0d716b01-a5d5-40f6-b002-5e1488bd54af" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.442329] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Lock "0d716b01-a5d5-40f6-b002-5e1488bd54af" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.874132] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276310, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.849257} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.874577] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Copied Virtual Disk [datastore1] vmware_temp/26e44410-c72a-474e-b6fc-16d791982618/4ca02567-c128-482f-b204-6ac166ac3160/tmp-sparse.vmdk to [datastore1] vmware_temp/26e44410-c72a-474e-b6fc-16d791982618/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 537.874577] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Deleting the datastore file [datastore1] vmware_temp/26e44410-c72a-474e-b6fc-16d791982618/4ca02567-c128-482f-b204-6ac166ac3160/tmp-sparse.vmdk {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 537.875932] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f7be0fae-f991-4079-905c-d02cb1d71814 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.891572] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Waiting for the task: (returnval){ [ 537.891572] env[61867]: value = "task-1276311" [ 537.891572] env[61867]: _type = "Task" [ 537.891572] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.903445] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276311, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.126999] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99cd450f-0e33-4c89-94ac-2ff5f10a5bd6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.134745] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75b612ec-2ed4-46de-a5c7-ce272aefeca4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.167958] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6df3b35-f880-45bb-aab3-229c990c4c6d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.176610] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a98d18-fb62-4614-9ead-50d0065be648 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.194410] env[61867]: DEBUG nova.compute.provider_tree [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 538.394800] env[61867]: DEBUG nova.compute.manager [req-43bd274c-0458-453f-a9cf-d46092909a24 req-0a41e100-951f-42e7-bbc0-1ae7486883a1 service nova] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Received event network-changed-b3b1de3c-67f3-4a32-a601-90f3081ce672 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 538.395236] env[61867]: DEBUG nova.compute.manager [req-43bd274c-0458-453f-a9cf-d46092909a24 req-0a41e100-951f-42e7-bbc0-1ae7486883a1 service nova] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Refreshing instance network info cache due to event network-changed-b3b1de3c-67f3-4a32-a601-90f3081ce672. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 538.395236] env[61867]: DEBUG oslo_concurrency.lockutils [req-43bd274c-0458-453f-a9cf-d46092909a24 req-0a41e100-951f-42e7-bbc0-1ae7486883a1 service nova] Acquiring lock "refresh_cache-ac5df149-83db-481a-a240-e5cf08fdbff9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.395380] env[61867]: DEBUG oslo_concurrency.lockutils [req-43bd274c-0458-453f-a9cf-d46092909a24 req-0a41e100-951f-42e7-bbc0-1ae7486883a1 service nova] Acquired lock "refresh_cache-ac5df149-83db-481a-a240-e5cf08fdbff9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.395472] env[61867]: DEBUG nova.network.neutron [req-43bd274c-0458-453f-a9cf-d46092909a24 req-0a41e100-951f-42e7-bbc0-1ae7486883a1 service nova] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Refreshing network info cache for port b3b1de3c-67f3-4a32-a601-90f3081ce672 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 538.409352] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276311, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.028327} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.409601] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 538.409804] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Moving file from [datastore1] vmware_temp/26e44410-c72a-474e-b6fc-16d791982618/4ca02567-c128-482f-b204-6ac166ac3160 to [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160. {{(pid=61867) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 538.410065] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-3c710ce0-937d-46a0-b856-e770094e2575 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.420195] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Waiting for the task: (returnval){ [ 538.420195] env[61867]: value = "task-1276312" [ 538.420195] env[61867]: _type = "Task" [ 538.420195] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.432636] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276312, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.543323] env[61867]: INFO nova.compute.manager [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Rebuilding instance [ 538.590979] env[61867]: DEBUG nova.compute.manager [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 538.591922] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf67f617-4774-4d6c-9261-e7f90f884d4d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.699011] env[61867]: DEBUG nova.scheduler.client.report [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 538.931053] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276312, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.049315} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.931346] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] File moved {{(pid=61867) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 538.931647] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Cleaning up location [datastore1] vmware_temp/26e44410-c72a-474e-b6fc-16d791982618 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 538.931713] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Deleting the datastore file [datastore1] vmware_temp/26e44410-c72a-474e-b6fc-16d791982618 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 538.931966] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-25866a23-ab2d-4b19-98e6-b9faafd055b5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.940711] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Waiting for the task: (returnval){ [ 538.940711] env[61867]: value = "task-1276313" [ 538.940711] env[61867]: _type = "Task" [ 538.940711] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.950897] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276313, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.969926] env[61867]: DEBUG nova.network.neutron [req-43bd274c-0458-453f-a9cf-d46092909a24 req-0a41e100-951f-42e7-bbc0-1ae7486883a1 service nova] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.104024] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 539.104513] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f7da0ce-5f2d-476e-8e6d-55a12a50b28f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.113217] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Waiting for the task: (returnval){ [ 539.113217] env[61867]: value = "task-1276314" [ 539.113217] env[61867]: _type = "Task" [ 539.113217] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.123226] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': task-1276314, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.205425] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.205962] env[61867]: DEBUG nova.compute.manager [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 539.211389] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.509s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.338186] env[61867]: DEBUG nova.network.neutron [req-43bd274c-0458-453f-a9cf-d46092909a24 req-0a41e100-951f-42e7-bbc0-1ae7486883a1 service nova] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.455285] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276313, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.031221} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.455555] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 539.459462] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f565f0da-3118-4da0-8c98-a985f572e555 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.462992] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Waiting for the task: (returnval){ [ 539.462992] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5252f6c5-a410-f317-4204-1a4b4dc93e46" [ 539.462992] env[61867]: _type = "Task" [ 539.462992] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.472200] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5252f6c5-a410-f317-4204-1a4b4dc93e46, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.626013] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': task-1276314, 'name': PowerOffVM_Task, 'duration_secs': 0.109643} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.626539] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 539.626894] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 539.627850] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d934367-f614-4494-967f-43915c019803 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.635754] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 539.636055] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c7b590d9-0135-41b3-a266-7da5565e4e0c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.659765] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 539.659983] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 539.660180] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Deleting the datastore file [datastore2] 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 539.660438] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9033b7d8-f265-4117-8a87-5045724c612c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.667437] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Waiting for the task: (returnval){ [ 539.667437] env[61867]: value = "task-1276316" [ 539.667437] env[61867]: _type = "Task" [ 539.667437] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.676045] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': task-1276316, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.716620] env[61867]: DEBUG nova.compute.utils [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 539.718395] env[61867]: DEBUG nova.compute.manager [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 539.718395] env[61867]: DEBUG nova.network.neutron [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 539.829710] env[61867]: DEBUG nova.policy [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc358009ebe6495a881034439d00978f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a29c5e9cfdaa4cc88ed300100a308ffa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 539.840153] env[61867]: DEBUG oslo_concurrency.lockutils [req-43bd274c-0458-453f-a9cf-d46092909a24 req-0a41e100-951f-42e7-bbc0-1ae7486883a1 service nova] Releasing lock "refresh_cache-ac5df149-83db-481a-a240-e5cf08fdbff9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.840153] env[61867]: DEBUG nova.compute.manager [req-43bd274c-0458-453f-a9cf-d46092909a24 req-0a41e100-951f-42e7-bbc0-1ae7486883a1 service nova] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Received event network-vif-deleted-b3b1de3c-67f3-4a32-a601-90f3081ce672 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.977253] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5252f6c5-a410-f317-4204-1a4b4dc93e46, 'name': SearchDatastore_Task, 'duration_secs': 0.008688} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.979169] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.979475] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 76d3a228-e314-42b1-9eec-43a32fbd8035/76d3a228-e314-42b1-9eec-43a32fbd8035.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 539.980333] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc6a0a2-a0f7-416c-bd05-eda4a21d3c8b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.982909] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3d2193f5-5f2c-47c8-9d7f-065bec6237ca {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.993987] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610f1950-68fd-4b17-8082-bd432f206d23 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.993987] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Waiting for the task: (returnval){ [ 539.993987] env[61867]: value = "task-1276317" [ 539.993987] env[61867]: _type = "Task" [ 539.993987] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.033233] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a246ca7-83a7-4709-bd3e-f990daf42fd2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.038793] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276317, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.044024] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f54ee83-496e-4270-8f1d-06df7b4b78b6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.057810] env[61867]: DEBUG nova.compute.provider_tree [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 540.183676] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': task-1276316, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.086417} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.184082] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 540.184482] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 540.184826] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 540.222701] env[61867]: DEBUG nova.compute.manager [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 540.508337] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276317, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49165} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.508866] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 76d3a228-e314-42b1-9eec-43a32fbd8035/76d3a228-e314-42b1-9eec-43a32fbd8035.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 540.509147] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 540.509451] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d58edcba-6de7-49a3-bc8f-26e2f4eca350 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.516301] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Waiting for the task: (returnval){ [ 540.516301] env[61867]: value = "task-1276318" [ 540.516301] env[61867]: _type = "Task" [ 540.516301] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.525203] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276318, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.561059] env[61867]: DEBUG nova.scheduler.client.report [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 540.592580] env[61867]: DEBUG nova.network.neutron [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Successfully created port: b2e60b2d-00eb-419a-b4c7-7c83e016cd65 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 541.026623] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276318, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060931} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.027054] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 541.027880] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-debd6078-448c-4423-820f-a0f28307bb32 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.052835] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Reconfiguring VM instance instance-00000009 to attach disk [datastore1] 76d3a228-e314-42b1-9eec-43a32fbd8035/76d3a228-e314-42b1-9eec-43a32fbd8035.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 541.053570] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7651b94e-4cb5-46d2-8e15-8bbe23beb12a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.068192] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.857s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.068857] env[61867]: ERROR nova.compute.manager [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e0651f92-1ada-45ee-b225-f8bf6b977110, please check neutron logs for more information. [ 541.068857] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Traceback (most recent call last): [ 541.068857] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 541.068857] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] self.driver.spawn(context, instance, image_meta, [ 541.068857] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 541.068857] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 541.068857] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 541.068857] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] vm_ref = self.build_virtual_machine(instance, [ 541.068857] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 541.068857] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] vif_infos = vmwarevif.get_vif_info(self._session, [ 541.068857] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 541.069239] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] for vif in network_info: [ 541.069239] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 541.069239] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] return self._sync_wrapper(fn, *args, **kwargs) [ 541.069239] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 541.069239] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] self.wait() [ 541.069239] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 541.069239] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] self[:] = self._gt.wait() [ 541.069239] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 541.069239] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] return self._exit_event.wait() [ 541.069239] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 541.069239] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] result = hub.switch() [ 541.069239] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 541.069239] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] return self.greenlet.switch() [ 541.069592] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.069592] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] result = function(*args, **kwargs) [ 541.069592] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 541.069592] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] return func(*args, **kwargs) [ 541.069592] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.069592] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] raise e [ 541.069592] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.069592] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] nwinfo = self.network_api.allocate_for_instance( [ 541.069592] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.069592] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] created_port_ids = self._update_ports_for_instance( [ 541.069592] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.069592] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] with excutils.save_and_reraise_exception(): [ 541.069592] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.069940] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] self.force_reraise() [ 541.069940] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.069940] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] raise self.value [ 541.069940] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.069940] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] updated_port = self._update_port( [ 541.069940] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.069940] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] _ensure_no_port_binding_failure(port) [ 541.069940] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.069940] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] raise exception.PortBindingFailed(port_id=port['id']) [ 541.069940] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] nova.exception.PortBindingFailed: Binding failed for port e0651f92-1ada-45ee-b225-f8bf6b977110, please check neutron logs for more information. [ 541.069940] env[61867]: ERROR nova.compute.manager [instance: dad6d190-9e74-452b-bb76-6d62f283218e] [ 541.073018] env[61867]: DEBUG nova.compute.utils [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Binding failed for port e0651f92-1ada-45ee-b225-f8bf6b977110, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 541.073018] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.409s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.074028] env[61867]: INFO nova.compute.claims [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 541.078167] env[61867]: DEBUG nova.compute.manager [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Build of instance dad6d190-9e74-452b-bb76-6d62f283218e was re-scheduled: Binding failed for port e0651f92-1ada-45ee-b225-f8bf6b977110, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 541.078651] env[61867]: DEBUG nova.compute.manager [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 541.078899] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Acquiring lock "refresh_cache-dad6d190-9e74-452b-bb76-6d62f283218e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.079090] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Acquired lock "refresh_cache-dad6d190-9e74-452b-bb76-6d62f283218e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.079278] env[61867]: DEBUG nova.network.neutron [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 541.082467] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Waiting for the task: (returnval){ [ 541.082467] env[61867]: value = "task-1276319" [ 541.082467] env[61867]: _type = "Task" [ 541.082467] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.090267] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276319, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.234112] env[61867]: DEBUG nova.virt.hardware [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 541.234490] env[61867]: DEBUG nova.virt.hardware [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 541.234530] env[61867]: DEBUG nova.virt.hardware [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 541.234740] env[61867]: DEBUG nova.virt.hardware [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 541.234986] env[61867]: DEBUG nova.virt.hardware [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 541.235542] env[61867]: DEBUG nova.virt.hardware [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 541.235806] env[61867]: DEBUG nova.virt.hardware [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 541.235985] env[61867]: DEBUG nova.virt.hardware [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 541.236170] env[61867]: DEBUG nova.virt.hardware [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 541.236335] env[61867]: DEBUG nova.virt.hardware [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 541.236524] env[61867]: DEBUG nova.virt.hardware [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 541.237879] env[61867]: DEBUG nova.compute.manager [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 541.242789] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb61435-7006-407e-832f-c6bc7c6d533f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.257036] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4219fa4d-aa8f-4a75-a5fd-d0db42eb9911 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.283176] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Instance VIF info [] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 541.289382] env[61867]: DEBUG oslo.service.loopingcall [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 541.294860] env[61867]: DEBUG nova.virt.hardware [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 541.295893] env[61867]: DEBUG nova.virt.hardware [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 541.298446] env[61867]: DEBUG nova.virt.hardware [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 541.298639] env[61867]: DEBUG nova.virt.hardware [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 541.298800] env[61867]: DEBUG nova.virt.hardware [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 541.298951] env[61867]: DEBUG nova.virt.hardware [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 541.299183] env[61867]: DEBUG nova.virt.hardware [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 541.299341] env[61867]: DEBUG nova.virt.hardware [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 541.299507] env[61867]: DEBUG nova.virt.hardware [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 541.299723] env[61867]: DEBUG nova.virt.hardware [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 541.299832] env[61867]: DEBUG nova.virt.hardware [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 541.300153] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 541.300915] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46347071-8602-4c15-9736-2de246dcdf62 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.304168] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7818bdb5-3d60-4be3-8c2c-66271501badb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.329427] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fdbe7c3-59a1-484c-9283-cd6331348414 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.338527] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 541.338527] env[61867]: value = "task-1276320" [ 541.338527] env[61867]: _type = "Task" [ 541.338527] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.354473] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276320, 'name': CreateVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.597984] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276319, 'name': ReconfigVM_Task, 'duration_secs': 0.279757} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.598286] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Reconfigured VM instance instance-00000009 to attach disk [datastore1] 76d3a228-e314-42b1-9eec-43a32fbd8035/76d3a228-e314-42b1-9eec-43a32fbd8035.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 541.598886] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ab725482-6914-406e-aede-5d000deacb2f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.606266] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Waiting for the task: (returnval){ [ 541.606266] env[61867]: value = "task-1276321" [ 541.606266] env[61867]: _type = "Task" [ 541.606266] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.616524] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276321, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.619850] env[61867]: DEBUG nova.network.neutron [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 541.850897] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276320, 'name': CreateVM_Task, 'duration_secs': 0.318025} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.851090] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 541.852116] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.852116] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.852116] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 541.852258] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a26bfd8a-6b84-49bb-99a3-4c9b685dad22 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.858804] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Waiting for the task: (returnval){ [ 541.858804] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52fce96b-0313-1d31-1a22-d2dadef2db23" [ 541.858804] env[61867]: _type = "Task" [ 541.858804] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.868545] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52fce96b-0313-1d31-1a22-d2dadef2db23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.917157] env[61867]: DEBUG nova.network.neutron [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.118015] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276321, 'name': Rename_Task, 'duration_secs': 0.129118} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.120676] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 542.121112] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ea1a2b4-bd4d-4ff6-92c1-fb3fb1f3cb94 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.128494] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Waiting for the task: (returnval){ [ 542.128494] env[61867]: value = "task-1276322" [ 542.128494] env[61867]: _type = "Task" [ 542.128494] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.143890] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276322, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.273310] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquiring lock "76c297c0-f393-499a-80e5-a78d84ec0885" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.274199] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "76c297c0-f393-499a-80e5-a78d84ec0885" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.356183] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd2310c-eee6-4d9e-bf6d-db97c14911d0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.373665] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cad38a4f-1f41-4556-86b2-449a10cb6f90 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.375391] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52fce96b-0313-1d31-1a22-d2dadef2db23, 'name': SearchDatastore_Task, 'duration_secs': 0.008552} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.376057] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.376368] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 542.376907] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.377069] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.377308] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 542.378173] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8b78f9b7-c1b0-4299-b8a0-b21284187192 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.407764] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b362c6-10b8-4644-a9c3-801a1865a04b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.416522] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 542.416522] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 542.417327] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e66237b7-2721-4432-bee7-b90e0dfc6672 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.420060] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Releasing lock "refresh_cache-dad6d190-9e74-452b-bb76-6d62f283218e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.420328] env[61867]: DEBUG nova.compute.manager [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 542.420554] env[61867]: DEBUG nova.compute.manager [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 542.420764] env[61867]: DEBUG nova.network.neutron [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 542.423878] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a192cf-9a65-424b-88f3-b34ec2e1aaae {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.430398] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Waiting for the task: (returnval){ [ 542.430398] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52c649c7-4562-5853-c4ae-e9535b66eb87" [ 542.430398] env[61867]: _type = "Task" [ 542.430398] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.439820] env[61867]: DEBUG nova.compute.provider_tree [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 542.453118] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52c649c7-4562-5853-c4ae-e9535b66eb87, 'name': SearchDatastore_Task, 'duration_secs': 0.012515} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.453118] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e21c3838-5757-4ed4-a444-7b4c3e11bd28 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.456561] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Waiting for the task: (returnval){ [ 542.456561] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52daff7a-1064-2bb4-208b-ee162b472521" [ 542.456561] env[61867]: _type = "Task" [ 542.456561] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.465267] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52daff7a-1064-2bb4-208b-ee162b472521, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.476662] env[61867]: DEBUG nova.network.neutron [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.645980] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276322, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.946685] env[61867]: DEBUG nova.scheduler.client.report [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 542.975230] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52daff7a-1064-2bb4-208b-ee162b472521, 'name': SearchDatastore_Task, 'duration_secs': 0.00847} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.976231] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.976576] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa/318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 542.976814] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5d904247-234e-4e8e-a1a7-ec94216268e1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.979903] env[61867]: DEBUG nova.network.neutron [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.989265] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Waiting for the task: (returnval){ [ 542.989265] env[61867]: value = "task-1276323" [ 542.989265] env[61867]: _type = "Task" [ 542.989265] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.998115] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': task-1276323, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.144054] env[61867]: DEBUG oslo_vmware.api [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276322, 'name': PowerOnVM_Task, 'duration_secs': 0.598235} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.144102] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 543.144704] env[61867]: INFO nova.compute.manager [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Took 9.22 seconds to spawn the instance on the hypervisor. [ 543.144704] env[61867]: DEBUG nova.compute.manager [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 543.145384] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5769b086-500c-431e-9691-1de13910290d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.402180] env[61867]: ERROR nova.compute.manager [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b2e60b2d-00eb-419a-b4c7-7c83e016cd65, please check neutron logs for more information. [ 543.402180] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 543.402180] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.402180] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 543.402180] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.402180] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 543.402180] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.402180] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 543.402180] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.402180] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 543.402180] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.402180] env[61867]: ERROR nova.compute.manager raise self.value [ 543.402180] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.402180] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 543.402180] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.402180] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 543.402914] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.402914] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 543.402914] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b2e60b2d-00eb-419a-b4c7-7c83e016cd65, please check neutron logs for more information. [ 543.402914] env[61867]: ERROR nova.compute.manager [ 543.402914] env[61867]: Traceback (most recent call last): [ 543.402914] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 543.402914] env[61867]: listener.cb(fileno) [ 543.402914] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.402914] env[61867]: result = function(*args, **kwargs) [ 543.402914] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 543.402914] env[61867]: return func(*args, **kwargs) [ 543.402914] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.402914] env[61867]: raise e [ 543.402914] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.402914] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 543.402914] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.402914] env[61867]: created_port_ids = self._update_ports_for_instance( [ 543.402914] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.402914] env[61867]: with excutils.save_and_reraise_exception(): [ 543.402914] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.402914] env[61867]: self.force_reraise() [ 543.402914] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.402914] env[61867]: raise self.value [ 543.402914] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.402914] env[61867]: updated_port = self._update_port( [ 543.402914] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.402914] env[61867]: _ensure_no_port_binding_failure(port) [ 543.402914] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.402914] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 543.403744] env[61867]: nova.exception.PortBindingFailed: Binding failed for port b2e60b2d-00eb-419a-b4c7-7c83e016cd65, please check neutron logs for more information. [ 543.403744] env[61867]: Removing descriptor: 17 [ 543.403744] env[61867]: ERROR nova.compute.manager [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b2e60b2d-00eb-419a-b4c7-7c83e016cd65, please check neutron logs for more information. [ 543.403744] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Traceback (most recent call last): [ 543.403744] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 543.403744] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] yield resources [ 543.403744] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 543.403744] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] self.driver.spawn(context, instance, image_meta, [ 543.403744] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 543.403744] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 543.403744] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 543.403744] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] vm_ref = self.build_virtual_machine(instance, [ 543.404205] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 543.404205] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] vif_infos = vmwarevif.get_vif_info(self._session, [ 543.404205] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 543.404205] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] for vif in network_info: [ 543.404205] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 543.404205] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] return self._sync_wrapper(fn, *args, **kwargs) [ 543.404205] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 543.404205] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] self.wait() [ 543.404205] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 543.404205] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] self[:] = self._gt.wait() [ 543.404205] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 543.404205] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] return self._exit_event.wait() [ 543.404205] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 543.404556] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] result = hub.switch() [ 543.404556] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 543.404556] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] return self.greenlet.switch() [ 543.404556] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.404556] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] result = function(*args, **kwargs) [ 543.404556] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 543.404556] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] return func(*args, **kwargs) [ 543.404556] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.404556] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] raise e [ 543.404556] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.404556] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] nwinfo = self.network_api.allocate_for_instance( [ 543.404556] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.404556] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] created_port_ids = self._update_ports_for_instance( [ 543.404942] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.404942] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] with excutils.save_and_reraise_exception(): [ 543.404942] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.404942] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] self.force_reraise() [ 543.404942] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.404942] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] raise self.value [ 543.404942] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.404942] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] updated_port = self._update_port( [ 543.404942] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.404942] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] _ensure_no_port_binding_failure(port) [ 543.404942] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.404942] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] raise exception.PortBindingFailed(port_id=port['id']) [ 543.405321] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] nova.exception.PortBindingFailed: Binding failed for port b2e60b2d-00eb-419a-b4c7-7c83e016cd65, please check neutron logs for more information. [ 543.405321] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] [ 543.405321] env[61867]: INFO nova.compute.manager [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Terminating instance [ 543.406385] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "refresh_cache-a746fcff-646e-4f9b-b595-7379c6f97efd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.406605] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquired lock "refresh_cache-a746fcff-646e-4f9b-b595-7379c6f97efd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.406790] env[61867]: DEBUG nova.network.neutron [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 543.450580] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.379s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.451145] env[61867]: DEBUG nova.compute.manager [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 543.455063] env[61867]: DEBUG oslo_concurrency.lockutils [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.542s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.456740] env[61867]: INFO nova.compute.claims [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 543.484023] env[61867]: INFO nova.compute.manager [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] Took 1.06 seconds to deallocate network for instance. [ 543.504650] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': task-1276323, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.669599] env[61867]: INFO nova.compute.manager [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Took 18.43 seconds to build instance. [ 543.855732] env[61867]: DEBUG nova.compute.manager [req-e5a52136-ea1a-4820-bc5f-5552e10b16ec req-bbabdc74-b9f9-4f67-8b81-a7adcc9d155d service nova] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Received event network-changed-b2e60b2d-00eb-419a-b4c7-7c83e016cd65 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 543.855732] env[61867]: DEBUG nova.compute.manager [req-e5a52136-ea1a-4820-bc5f-5552e10b16ec req-bbabdc74-b9f9-4f67-8b81-a7adcc9d155d service nova] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Refreshing instance network info cache due to event network-changed-b2e60b2d-00eb-419a-b4c7-7c83e016cd65. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 543.855732] env[61867]: DEBUG oslo_concurrency.lockutils [req-e5a52136-ea1a-4820-bc5f-5552e10b16ec req-bbabdc74-b9f9-4f67-8b81-a7adcc9d155d service nova] Acquiring lock "refresh_cache-a746fcff-646e-4f9b-b595-7379c6f97efd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.967370] env[61867]: DEBUG nova.compute.utils [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 543.969718] env[61867]: DEBUG nova.network.neutron [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.979864] env[61867]: DEBUG nova.compute.manager [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 543.979864] env[61867]: DEBUG nova.network.neutron [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 544.009398] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': task-1276323, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.845036} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.009398] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa/318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 544.009398] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 544.009654] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-21815ad4-3fd1-46ec-b254-382b8b33e9fe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.021122] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Waiting for the task: (returnval){ [ 544.021122] env[61867]: value = "task-1276324" [ 544.021122] env[61867]: _type = "Task" [ 544.021122] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.035257] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': task-1276324, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.170074] env[61867]: DEBUG nova.policy [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '71ac042083dd4903bf774af25198bfc7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ab9ced31f8240debabd53e19e50d32a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 544.173644] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2eac48b1-693d-4da1-bfec-3c75b4f32c21 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Lock "76d3a228-e314-42b1-9eec-43a32fbd8035" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.943s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.350492] env[61867]: DEBUG nova.network.neutron [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.483054] env[61867]: DEBUG nova.compute.manager [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 544.536296] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': task-1276324, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073742} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.539384] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 544.540258] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a049ef4d-710f-439d-a30e-e11592731984 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.555231] env[61867]: INFO nova.scheduler.client.report [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Deleted allocations for instance dad6d190-9e74-452b-bb76-6d62f283218e [ 544.581819] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa/318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 544.588216] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f8fc644-4357-4c77-a8fb-3a1f6b7f5495 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.610401] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Waiting for the task: (returnval){ [ 544.610401] env[61867]: value = "task-1276325" [ 544.610401] env[61867]: _type = "Task" [ 544.610401] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.627257] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': task-1276325, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.670576] env[61867]: DEBUG oslo_concurrency.lockutils [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquiring lock "af4b928f-25bc-4eff-9ba9-58afeae887f2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.671050] env[61867]: DEBUG oslo_concurrency.lockutils [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "af4b928f-25bc-4eff-9ba9-58afeae887f2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.681627] env[61867]: DEBUG nova.compute.manager [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 544.820435] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d2f6c2-bf90-4cdf-9c07-5f09e68976ca {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.829205] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc2fb3d-7ea1-4b49-8379-8413fd6bf226 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.868775] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Releasing lock "refresh_cache-a746fcff-646e-4f9b-b595-7379c6f97efd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.869292] env[61867]: DEBUG nova.compute.manager [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 544.869554] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 544.870476] env[61867]: DEBUG oslo_concurrency.lockutils [req-e5a52136-ea1a-4820-bc5f-5552e10b16ec req-bbabdc74-b9f9-4f67-8b81-a7adcc9d155d service nova] Acquired lock "refresh_cache-a746fcff-646e-4f9b-b595-7379c6f97efd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.870821] env[61867]: DEBUG nova.network.neutron [req-e5a52136-ea1a-4820-bc5f-5552e10b16ec req-bbabdc74-b9f9-4f67-8b81-a7adcc9d155d service nova] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Refreshing network info cache for port b2e60b2d-00eb-419a-b4c7-7c83e016cd65 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 544.872090] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f8fba37b-c870-4329-8756-2f7d385bf9e7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.874718] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d8b488-fd2e-4fa2-8c70-355488cc9256 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.885997] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-885df289-8059-4d4c-aa81-f91912635b33 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.895459] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af939e9e-0019-434e-804b-871add15862d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.917491] env[61867]: DEBUG nova.compute.provider_tree [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 544.923936] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a746fcff-646e-4f9b-b595-7379c6f97efd could not be found. [ 544.924188] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 544.924367] env[61867]: INFO nova.compute.manager [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Took 0.05 seconds to destroy the instance on the hypervisor. [ 544.924649] env[61867]: DEBUG oslo.service.loopingcall [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 544.925144] env[61867]: DEBUG nova.compute.manager [-] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 544.925259] env[61867]: DEBUG nova.network.neutron [-] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 544.969701] env[61867]: DEBUG nova.network.neutron [-] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.056924] env[61867]: DEBUG nova.compute.manager [None req-a5f7a664-e6bc-47bc-a3c2-0dcdbb8ab42b tempest-ServerDiagnosticsV248Test-1182926949 tempest-ServerDiagnosticsV248Test-1182926949-project-admin] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 545.058162] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d755f0-9f7b-489a-bdd9-8a2a756917de {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.065914] env[61867]: INFO nova.compute.manager [None req-a5f7a664-e6bc-47bc-a3c2-0dcdbb8ab42b tempest-ServerDiagnosticsV248Test-1182926949 tempest-ServerDiagnosticsV248Test-1182926949-project-admin] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Retrieving diagnostics [ 545.066685] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e25fdc-2302-426a-ac8a-68b3d6bd5a73 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.105353] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7288955d-7042-496e-8112-98c991f4f43a tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Lock "dad6d190-9e74-452b-bb76-6d62f283218e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.053s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.105950] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "dad6d190-9e74-452b-bb76-6d62f283218e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 25.608s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.106164] env[61867]: INFO nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: dad6d190-9e74-452b-bb76-6d62f283218e] During sync_power_state the instance has a pending task (spawning). Skip. [ 545.106601] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "dad6d190-9e74-452b-bb76-6d62f283218e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.120296] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': task-1276325, 'name': ReconfigVM_Task, 'duration_secs': 0.278232} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.121206] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa/318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 545.121823] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-36a9e8ef-c3b7-4b8d-8f58-e3821e2d94ea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.129885] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Waiting for the task: (returnval){ [ 545.129885] env[61867]: value = "task-1276326" [ 545.129885] env[61867]: _type = "Task" [ 545.129885] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.145049] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': task-1276326, 'name': Rename_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.218249] env[61867]: DEBUG oslo_concurrency.lockutils [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.427207] env[61867]: DEBUG nova.scheduler.client.report [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 545.446128] env[61867]: DEBUG nova.network.neutron [req-e5a52136-ea1a-4820-bc5f-5552e10b16ec req-bbabdc74-b9f9-4f67-8b81-a7adcc9d155d service nova] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.473620] env[61867]: DEBUG nova.network.neutron [-] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.494080] env[61867]: DEBUG nova.compute.manager [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 545.497459] env[61867]: DEBUG nova.network.neutron [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Successfully created port: 770f79ea-89a4-44e5-ad5f-c7fcfb9ffa5b {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 545.537100] env[61867]: DEBUG nova.virt.hardware [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:43:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='495239252',id=24,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-319881821',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 545.537335] env[61867]: DEBUG nova.virt.hardware [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 545.537495] env[61867]: DEBUG nova.virt.hardware [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 545.537737] env[61867]: DEBUG nova.virt.hardware [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 545.537891] env[61867]: DEBUG nova.virt.hardware [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 545.541132] env[61867]: DEBUG nova.virt.hardware [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 545.541489] env[61867]: DEBUG nova.virt.hardware [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 545.541675] env[61867]: DEBUG nova.virt.hardware [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 545.541851] env[61867]: DEBUG nova.virt.hardware [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 545.542024] env[61867]: DEBUG nova.virt.hardware [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 545.542250] env[61867]: DEBUG nova.virt.hardware [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 545.543378] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb695cbb-6ea3-4575-80be-d04b8d4db71f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.552292] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0161c1a-811a-4a7a-b724-84f5bd656192 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.611216] env[61867]: DEBUG nova.compute.manager [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 545.647925] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': task-1276326, 'name': Rename_Task, 'duration_secs': 0.138139} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.648217] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 545.651928] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9315fff7-62ee-44d7-a44f-6517a6422e36 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.654646] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Waiting for the task: (returnval){ [ 545.654646] env[61867]: value = "task-1276327" [ 545.654646] env[61867]: _type = "Task" [ 545.654646] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.663753] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': task-1276327, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.844991] env[61867]: DEBUG nova.network.neutron [req-e5a52136-ea1a-4820-bc5f-5552e10b16ec req-bbabdc74-b9f9-4f67-8b81-a7adcc9d155d service nova] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.938237] env[61867]: DEBUG oslo_concurrency.lockutils [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.938656] env[61867]: DEBUG nova.compute.manager [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 545.942537] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.445s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.982694] env[61867]: INFO nova.compute.manager [-] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Took 1.06 seconds to deallocate network for instance. [ 545.985481] env[61867]: DEBUG nova.compute.claims [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 545.985578] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.130640] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquiring lock "45e2004d-731b-4002-84b2-3f9138c09663" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.130972] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "45e2004d-731b-4002-84b2-3f9138c09663" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.149634] env[61867]: DEBUG oslo_concurrency.lockutils [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.171737] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': task-1276327, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.348273] env[61867]: DEBUG oslo_concurrency.lockutils [req-e5a52136-ea1a-4820-bc5f-5552e10b16ec req-bbabdc74-b9f9-4f67-8b81-a7adcc9d155d service nova] Releasing lock "refresh_cache-a746fcff-646e-4f9b-b595-7379c6f97efd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.448347] env[61867]: DEBUG nova.compute.utils [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 546.452516] env[61867]: DEBUG nova.compute.manager [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 546.452516] env[61867]: DEBUG nova.network.neutron [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 546.605573] env[61867]: DEBUG nova.policy [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d745e09de684304acdb07dfe7aab62d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f05b6399ebfd4e4692a0e93bf1b9c7cb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 546.670815] env[61867]: DEBUG oslo_vmware.api [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Task: {'id': task-1276327, 'name': PowerOnVM_Task, 'duration_secs': 0.515947} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.671104] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 546.671344] env[61867]: DEBUG nova.compute.manager [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 546.672150] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-823ab70e-ee5f-49b9-b03d-311820b1fbfc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.712666] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7caaed-fe2c-4f69-ac04-5b3c7b9a3e57 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.731192] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4baa0e3-8753-42e9-931a-a82d047f20ba {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.776238] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f471cc5-cf65-4573-bb44-2410c12430ec {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.788067] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2577f59e-7648-45e1-ad35-6268666d08f4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.801793] env[61867]: DEBUG nova.compute.provider_tree [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 546.953466] env[61867]: DEBUG nova.compute.manager [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 547.198191] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.306690] env[61867]: DEBUG nova.scheduler.client.report [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 547.463746] env[61867]: INFO nova.virt.block_device [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Booting with volume 4ab7bf71-cd89-45f4-8d42-38aecc267234 at /dev/sda [ 547.471632] env[61867]: DEBUG nova.compute.manager [req-955aedd6-76ae-49d1-b316-c334f660117d req-d2c68acf-fd92-422f-a51d-d45a1074d52b service nova] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Received event network-vif-deleted-b2e60b2d-00eb-419a-b4c7-7c83e016cd65 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 547.537091] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b7527511-8e72-406c-b4cb-71c574162324 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.548256] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f123498-e2e4-4f7a-8c22-a81f3037741a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.571854] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-64430d1c-8e88-46b0-af94-f223ebaabe19 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.580907] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c787c1-abf1-46a4-a166-dda2c1949b76 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.610742] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bdf98d7-439a-4c1e-b19d-8d49305f402f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.617899] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44aa797f-2200-4fe5-8d92-0c5b5be501fd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.631959] env[61867]: DEBUG nova.virt.block_device [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Updating existing volume attachment record: 4e3bb226-8193-4895-9ee0-7f2d3a216aac {{(pid=61867) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 547.810904] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.868s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.811583] env[61867]: ERROR nova.compute.manager [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f757b37f-f7bf-48b8-be0d-0ad5e3eadd6b, please check neutron logs for more information. [ 547.811583] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Traceback (most recent call last): [ 547.811583] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 547.811583] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] self.driver.spawn(context, instance, image_meta, [ 547.811583] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 547.811583] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.811583] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.811583] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] vm_ref = self.build_virtual_machine(instance, [ 547.811583] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.811583] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.811583] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.812176] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] for vif in network_info: [ 547.812176] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.812176] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] return self._sync_wrapper(fn, *args, **kwargs) [ 547.812176] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.812176] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] self.wait() [ 547.812176] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.812176] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] self[:] = self._gt.wait() [ 547.812176] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.812176] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] return self._exit_event.wait() [ 547.812176] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.812176] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] result = hub.switch() [ 547.812176] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.812176] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] return self.greenlet.switch() [ 547.812542] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.812542] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] result = function(*args, **kwargs) [ 547.812542] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.812542] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] return func(*args, **kwargs) [ 547.812542] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 547.812542] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] raise e [ 547.812542] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.812542] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] nwinfo = self.network_api.allocate_for_instance( [ 547.812542] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.812542] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] created_port_ids = self._update_ports_for_instance( [ 547.812542] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.812542] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] with excutils.save_and_reraise_exception(): [ 547.812542] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.812912] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] self.force_reraise() [ 547.812912] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.812912] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] raise self.value [ 547.812912] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.812912] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] updated_port = self._update_port( [ 547.812912] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.812912] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] _ensure_no_port_binding_failure(port) [ 547.812912] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.812912] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] raise exception.PortBindingFailed(port_id=port['id']) [ 547.812912] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] nova.exception.PortBindingFailed: Binding failed for port f757b37f-f7bf-48b8-be0d-0ad5e3eadd6b, please check neutron logs for more information. [ 547.812912] env[61867]: ERROR nova.compute.manager [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] [ 547.813242] env[61867]: DEBUG nova.compute.utils [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Binding failed for port f757b37f-f7bf-48b8-be0d-0ad5e3eadd6b, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 547.813564] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.034s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.815138] env[61867]: INFO nova.compute.claims [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 547.817984] env[61867]: DEBUG nova.compute.manager [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Build of instance b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7 was re-scheduled: Binding failed for port f757b37f-f7bf-48b8-be0d-0ad5e3eadd6b, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 547.818511] env[61867]: DEBUG nova.compute.manager [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 547.818664] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Acquiring lock "refresh_cache-b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.818813] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Acquired lock "refresh_cache-b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.818969] env[61867]: DEBUG nova.network.neutron [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 548.124635] env[61867]: DEBUG nova.network.neutron [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Successfully created port: 7f66cc82-1538-49a5-8543-6c9722020aba {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 548.383718] env[61867]: DEBUG nova.network.neutron [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.900226] env[61867]: DEBUG nova.network.neutron [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.187643] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquiring lock "318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.188023] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Lock "318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.188107] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquiring lock "318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.188293] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Lock "318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.189139] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Lock "318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.193517] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a3bc682-f3d5-4ca6-bcaa-2a5d1b57c6da {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.199433] env[61867]: INFO nova.compute.manager [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Terminating instance [ 549.202699] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquiring lock "refresh_cache-318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.202926] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquired lock "refresh_cache-318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.204058] env[61867]: DEBUG nova.network.neutron [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 549.207816] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e5348a-dd4c-4d0f-8da8-5c4a476c0fef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.245436] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-324e9afe-58e6-4a7a-ae02-bf36ca1db471 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.255572] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405cc055-f4e2-45ad-b13b-893411a20ba4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.271806] env[61867]: DEBUG nova.compute.provider_tree [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 549.404654] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Releasing lock "refresh_cache-b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.404858] env[61867]: DEBUG nova.compute.manager [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 549.405048] env[61867]: DEBUG nova.compute.manager [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 549.405213] env[61867]: DEBUG nova.network.neutron [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 549.454131] env[61867]: DEBUG nova.network.neutron [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.754521] env[61867]: DEBUG nova.network.neutron [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.775886] env[61867]: DEBUG nova.scheduler.client.report [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 549.794947] env[61867]: DEBUG nova.compute.manager [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 549.795518] env[61867]: DEBUG nova.virt.hardware [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 549.795725] env[61867]: DEBUG nova.virt.hardware [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 549.795868] env[61867]: DEBUG nova.virt.hardware [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 549.796889] env[61867]: DEBUG nova.virt.hardware [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 549.796889] env[61867]: DEBUG nova.virt.hardware [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 549.796889] env[61867]: DEBUG nova.virt.hardware [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 549.796889] env[61867]: DEBUG nova.virt.hardware [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 549.797405] env[61867]: DEBUG nova.virt.hardware [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 549.797440] env[61867]: DEBUG nova.virt.hardware [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 549.797588] env[61867]: DEBUG nova.virt.hardware [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 549.797860] env[61867]: DEBUG nova.virt.hardware [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 549.798716] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf8b57f-eed0-45ec-9b94-cb8a3003da5c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.809105] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d130e917-4999-4c50-8989-5b1a8046defb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.957023] env[61867]: DEBUG nova.network.neutron [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.072947] env[61867]: DEBUG nova.network.neutron [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.281793] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.468s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.284014] env[61867]: DEBUG nova.compute.manager [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 550.286884] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.346s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.464764] env[61867]: INFO nova.compute.manager [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] Took 1.05 seconds to deallocate network for instance. [ 550.557232] env[61867]: DEBUG nova.compute.manager [req-2d31090c-bf2b-444b-b4f8-e0905b269aa2 req-576a87b1-221d-4ce2-b25e-5ea7e1810b89 service nova] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Received event network-changed-770f79ea-89a4-44e5-ad5f-c7fcfb9ffa5b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 550.557346] env[61867]: DEBUG nova.compute.manager [req-2d31090c-bf2b-444b-b4f8-e0905b269aa2 req-576a87b1-221d-4ce2-b25e-5ea7e1810b89 service nova] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Refreshing instance network info cache due to event network-changed-770f79ea-89a4-44e5-ad5f-c7fcfb9ffa5b. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 550.558022] env[61867]: DEBUG oslo_concurrency.lockutils [req-2d31090c-bf2b-444b-b4f8-e0905b269aa2 req-576a87b1-221d-4ce2-b25e-5ea7e1810b89 service nova] Acquiring lock "refresh_cache-c3f0399f-543c-4ab7-8854-b5f7f012fdd2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.558022] env[61867]: DEBUG oslo_concurrency.lockutils [req-2d31090c-bf2b-444b-b4f8-e0905b269aa2 req-576a87b1-221d-4ce2-b25e-5ea7e1810b89 service nova] Acquired lock "refresh_cache-c3f0399f-543c-4ab7-8854-b5f7f012fdd2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.558022] env[61867]: DEBUG nova.network.neutron [req-2d31090c-bf2b-444b-b4f8-e0905b269aa2 req-576a87b1-221d-4ce2-b25e-5ea7e1810b89 service nova] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Refreshing network info cache for port 770f79ea-89a4-44e5-ad5f-c7fcfb9ffa5b {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 550.573162] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Releasing lock "refresh_cache-318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.573879] env[61867]: DEBUG nova.compute.manager [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 550.574641] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 550.576146] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5580f56d-5102-483f-8e92-2a6f8ae39201 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.589020] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 550.590318] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc18dc39-8dd7-450d-b8d8-f8a68864d5e7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.597725] env[61867]: DEBUG oslo_vmware.api [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 550.597725] env[61867]: value = "task-1276328" [ 550.597725] env[61867]: _type = "Task" [ 550.597725] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.611205] env[61867]: DEBUG oslo_vmware.api [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276328, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.792672] env[61867]: DEBUG nova.compute.utils [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 550.802156] env[61867]: DEBUG nova.compute.manager [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 550.802156] env[61867]: DEBUG nova.network.neutron [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 551.000486] env[61867]: DEBUG nova.policy [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '121671d22ab744b2a017fefe850baed6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b72ce3f83d7f46cd976e942b8c4f11f5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 551.068720] env[61867]: ERROR nova.compute.manager [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 770f79ea-89a4-44e5-ad5f-c7fcfb9ffa5b, please check neutron logs for more information. [ 551.068720] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 551.068720] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.068720] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 551.068720] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.068720] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 551.068720] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.068720] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 551.068720] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.068720] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 551.068720] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.068720] env[61867]: ERROR nova.compute.manager raise self.value [ 551.068720] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.068720] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 551.068720] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.068720] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 551.069277] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.069277] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 551.069277] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 770f79ea-89a4-44e5-ad5f-c7fcfb9ffa5b, please check neutron logs for more information. [ 551.069277] env[61867]: ERROR nova.compute.manager [ 551.069277] env[61867]: Traceback (most recent call last): [ 551.069277] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 551.069277] env[61867]: listener.cb(fileno) [ 551.069277] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.069277] env[61867]: result = function(*args, **kwargs) [ 551.069277] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.069277] env[61867]: return func(*args, **kwargs) [ 551.069277] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.069277] env[61867]: raise e [ 551.069277] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.069277] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 551.069277] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.069277] env[61867]: created_port_ids = self._update_ports_for_instance( [ 551.069277] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.069277] env[61867]: with excutils.save_and_reraise_exception(): [ 551.069277] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.069277] env[61867]: self.force_reraise() [ 551.069277] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.069277] env[61867]: raise self.value [ 551.069277] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.069277] env[61867]: updated_port = self._update_port( [ 551.069277] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.069277] env[61867]: _ensure_no_port_binding_failure(port) [ 551.069277] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.069277] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 551.070070] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 770f79ea-89a4-44e5-ad5f-c7fcfb9ffa5b, please check neutron logs for more information. [ 551.070070] env[61867]: Removing descriptor: 14 [ 551.071637] env[61867]: ERROR nova.compute.manager [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 770f79ea-89a4-44e5-ad5f-c7fcfb9ffa5b, please check neutron logs for more information. [ 551.071637] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Traceback (most recent call last): [ 551.071637] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 551.071637] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] yield resources [ 551.071637] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 551.071637] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] self.driver.spawn(context, instance, image_meta, [ 551.071637] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 551.071637] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.071637] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.071637] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] vm_ref = self.build_virtual_machine(instance, [ 551.071637] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.072127] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.072127] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.072127] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] for vif in network_info: [ 551.072127] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.072127] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] return self._sync_wrapper(fn, *args, **kwargs) [ 551.072127] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.072127] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] self.wait() [ 551.072127] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.072127] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] self[:] = self._gt.wait() [ 551.072127] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.072127] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] return self._exit_event.wait() [ 551.072127] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.072127] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] result = hub.switch() [ 551.072493] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.072493] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] return self.greenlet.switch() [ 551.072493] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.072493] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] result = function(*args, **kwargs) [ 551.072493] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.072493] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] return func(*args, **kwargs) [ 551.072493] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.072493] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] raise e [ 551.072493] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.072493] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] nwinfo = self.network_api.allocate_for_instance( [ 551.072493] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.072493] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] created_port_ids = self._update_ports_for_instance( [ 551.072493] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.072880] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] with excutils.save_and_reraise_exception(): [ 551.072880] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.072880] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] self.force_reraise() [ 551.072880] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.072880] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] raise self.value [ 551.072880] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.072880] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] updated_port = self._update_port( [ 551.072880] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.072880] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] _ensure_no_port_binding_failure(port) [ 551.072880] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.072880] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] raise exception.PortBindingFailed(port_id=port['id']) [ 551.072880] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] nova.exception.PortBindingFailed: Binding failed for port 770f79ea-89a4-44e5-ad5f-c7fcfb9ffa5b, please check neutron logs for more information. [ 551.072880] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] [ 551.073298] env[61867]: INFO nova.compute.manager [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Terminating instance [ 551.074824] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Acquiring lock "refresh_cache-c3f0399f-543c-4ab7-8854-b5f7f012fdd2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.113841] env[61867]: DEBUG oslo_vmware.api [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276328, 'name': PowerOffVM_Task, 'duration_secs': 0.184295} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.114564] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 551.114564] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 551.114564] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cee0af36-7b5e-4007-9fec-5c28644fb936 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.123092] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a114ead3-d1ed-48b9-b354-1d794bee2fd8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.134873] env[61867]: DEBUG nova.network.neutron [req-2d31090c-bf2b-444b-b4f8-e0905b269aa2 req-576a87b1-221d-4ce2-b25e-5ea7e1810b89 service nova] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.139276] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32fb3e05-e005-416d-b939-72b4e687cac7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.144074] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 551.144329] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 551.144419] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Deleting the datastore file [datastore2] 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 551.144951] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-97c5f50d-e72d-458e-b6d5-2ef4fdd5c5be {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.178530] env[61867]: DEBUG oslo_vmware.api [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for the task: (returnval){ [ 551.178530] env[61867]: value = "task-1276330" [ 551.178530] env[61867]: _type = "Task" [ 551.178530] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.181152] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68497b44-dc5d-4e81-a50e-b2f1689b9acc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.194951] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85384f45-bf4b-4379-a707-2d295a49dc2c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.198941] env[61867]: DEBUG oslo_vmware.api [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276330, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.209499] env[61867]: DEBUG nova.compute.provider_tree [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 551.308462] env[61867]: DEBUG nova.compute.manager [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 551.509320] env[61867]: INFO nova.scheduler.client.report [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Deleted allocations for instance b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7 [ 551.611664] env[61867]: DEBUG nova.network.neutron [req-2d31090c-bf2b-444b-b4f8-e0905b269aa2 req-576a87b1-221d-4ce2-b25e-5ea7e1810b89 service nova] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.697417] env[61867]: DEBUG oslo_vmware.api [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Task: {'id': task-1276330, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.088587} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.697687] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 551.697860] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 551.698039] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 551.698210] env[61867]: INFO nova.compute.manager [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Took 1.12 seconds to destroy the instance on the hypervisor. [ 551.698456] env[61867]: DEBUG oslo.service.loopingcall [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 551.698642] env[61867]: DEBUG nova.compute.manager [-] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 551.698729] env[61867]: DEBUG nova.network.neutron [-] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 551.712779] env[61867]: DEBUG nova.scheduler.client.report [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 551.751330] env[61867]: DEBUG nova.network.neutron [-] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.018143] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4fb3e249-7b06-48d8-a06d-a52dddfad5ee tempest-ServersAdminTestJSON-388707207 tempest-ServersAdminTestJSON-388707207-project-member] Lock "b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.059s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.019841] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 32.521s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.020327] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4df55ce-ea4e-4d8b-8aae-af9a58a68d2e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.037976] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba19585-0b57-4d74-86cf-b9c016525fc2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.119024] env[61867]: DEBUG oslo_concurrency.lockutils [req-2d31090c-bf2b-444b-b4f8-e0905b269aa2 req-576a87b1-221d-4ce2-b25e-5ea7e1810b89 service nova] Releasing lock "refresh_cache-c3f0399f-543c-4ab7-8854-b5f7f012fdd2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.119506] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Acquired lock "refresh_cache-c3f0399f-543c-4ab7-8854-b5f7f012fdd2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.119698] env[61867]: DEBUG nova.network.neutron [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 552.218256] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.931s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.218376] env[61867]: ERROR nova.compute.manager [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5bdbe391-5c75-46f4-8747-a36c9b236366, please check neutron logs for more information. [ 552.218376] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Traceback (most recent call last): [ 552.218376] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 552.218376] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] self.driver.spawn(context, instance, image_meta, [ 552.218376] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 552.218376] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 552.218376] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 552.218376] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] vm_ref = self.build_virtual_machine(instance, [ 552.218376] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 552.218376] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 552.218376] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 552.218733] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] for vif in network_info: [ 552.218733] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 552.218733] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] return self._sync_wrapper(fn, *args, **kwargs) [ 552.218733] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 552.218733] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] self.wait() [ 552.218733] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 552.218733] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] self[:] = self._gt.wait() [ 552.218733] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 552.218733] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] return self._exit_event.wait() [ 552.218733] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 552.218733] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] result = hub.switch() [ 552.218733] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 552.218733] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] return self.greenlet.switch() [ 552.219378] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 552.219378] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] result = function(*args, **kwargs) [ 552.219378] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 552.219378] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] return func(*args, **kwargs) [ 552.219378] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 552.219378] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] raise e [ 552.219378] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 552.219378] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] nwinfo = self.network_api.allocate_for_instance( [ 552.219378] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 552.219378] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] created_port_ids = self._update_ports_for_instance( [ 552.219378] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 552.219378] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] with excutils.save_and_reraise_exception(): [ 552.219378] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 552.219931] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] self.force_reraise() [ 552.219931] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 552.219931] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] raise self.value [ 552.219931] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 552.219931] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] updated_port = self._update_port( [ 552.219931] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 552.219931] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] _ensure_no_port_binding_failure(port) [ 552.219931] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 552.219931] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] raise exception.PortBindingFailed(port_id=port['id']) [ 552.219931] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] nova.exception.PortBindingFailed: Binding failed for port 5bdbe391-5c75-46f4-8747-a36c9b236366, please check neutron logs for more information. [ 552.219931] env[61867]: ERROR nova.compute.manager [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] [ 552.220355] env[61867]: DEBUG nova.compute.utils [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Binding failed for port 5bdbe391-5c75-46f4-8747-a36c9b236366, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 552.225228] env[61867]: DEBUG nova.compute.manager [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Build of instance 433ce856-b04b-4808-8f05-201ad28e18d0 was re-scheduled: Binding failed for port 5bdbe391-5c75-46f4-8747-a36c9b236366, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 552.225228] env[61867]: DEBUG nova.compute.manager [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 552.225228] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Acquiring lock "refresh_cache-433ce856-b04b-4808-8f05-201ad28e18d0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.225228] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Acquired lock "refresh_cache-433ce856-b04b-4808-8f05-201ad28e18d0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.225754] env[61867]: DEBUG nova.network.neutron [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 552.225754] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.076s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.229336] env[61867]: INFO nova.compute.claims [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 552.253801] env[61867]: DEBUG nova.network.neutron [-] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.327280] env[61867]: DEBUG nova.compute.manager [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 552.367933] env[61867]: DEBUG nova.virt.hardware [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 552.368184] env[61867]: DEBUG nova.virt.hardware [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 552.368343] env[61867]: DEBUG nova.virt.hardware [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 552.368521] env[61867]: DEBUG nova.virt.hardware [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 552.368668] env[61867]: DEBUG nova.virt.hardware [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 552.368818] env[61867]: DEBUG nova.virt.hardware [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 552.369155] env[61867]: DEBUG nova.virt.hardware [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 552.369972] env[61867]: DEBUG nova.virt.hardware [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 552.369972] env[61867]: DEBUG nova.virt.hardware [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 552.369972] env[61867]: DEBUG nova.virt.hardware [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 552.369972] env[61867]: DEBUG nova.virt.hardware [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 552.370863] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5bfe77f-63e1-48d6-9ac2-16c973f68a91 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.380731] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0394d4e6-ea92-42a4-8142-68368751b239 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.473820] env[61867]: DEBUG nova.network.neutron [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Successfully created port: 59fcbb3a-4369-427e-ab62-f65f5db2eb86 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 552.528192] env[61867]: DEBUG nova.compute.manager [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 552.574385] env[61867]: INFO nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7] During the sync_power process the instance has moved from host None to host cpu-1 [ 552.575420] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "b747d6e6-1b07-4a7d-93ea-e8bb2b74d2f7" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.556s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.756336] env[61867]: INFO nova.compute.manager [-] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Took 1.06 seconds to deallocate network for instance. [ 552.783242] env[61867]: DEBUG nova.network.neutron [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.796954] env[61867]: DEBUG nova.network.neutron [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.807143] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "2ae534f8-1921-4892-8556-e2732624f412" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.807295] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "2ae534f8-1921-4892-8556-e2732624f412" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.833260] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "18b98634-d76d-496b-904f-1a9c289cdb8a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.834012] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "18b98634-d76d-496b-904f-1a9c289cdb8a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.054188] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.268057] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.291538] env[61867]: DEBUG nova.network.neutron [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.376994] env[61867]: DEBUG nova.network.neutron [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.519347] env[61867]: ERROR nova.compute.manager [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7f66cc82-1538-49a5-8543-6c9722020aba, please check neutron logs for more information. [ 553.519347] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 553.519347] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.519347] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 553.519347] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.519347] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 553.519347] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.519347] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 553.519347] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.519347] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 553.519347] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.519347] env[61867]: ERROR nova.compute.manager raise self.value [ 553.519347] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.519347] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 553.519347] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.519347] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 553.519880] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.519880] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 553.519880] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7f66cc82-1538-49a5-8543-6c9722020aba, please check neutron logs for more information. [ 553.519880] env[61867]: ERROR nova.compute.manager [ 553.519880] env[61867]: Traceback (most recent call last): [ 553.519880] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 553.519880] env[61867]: listener.cb(fileno) [ 553.519880] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.519880] env[61867]: result = function(*args, **kwargs) [ 553.519880] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 553.519880] env[61867]: return func(*args, **kwargs) [ 553.519880] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 553.519880] env[61867]: raise e [ 553.519880] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.519880] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 553.519880] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.519880] env[61867]: created_port_ids = self._update_ports_for_instance( [ 553.519880] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.519880] env[61867]: with excutils.save_and_reraise_exception(): [ 553.519880] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.519880] env[61867]: self.force_reraise() [ 553.519880] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.519880] env[61867]: raise self.value [ 553.519880] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.519880] env[61867]: updated_port = self._update_port( [ 553.519880] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.519880] env[61867]: _ensure_no_port_binding_failure(port) [ 553.519880] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.519880] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 553.520992] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 7f66cc82-1538-49a5-8543-6c9722020aba, please check neutron logs for more information. [ 553.520992] env[61867]: Removing descriptor: 17 [ 553.520992] env[61867]: ERROR nova.compute.manager [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7f66cc82-1538-49a5-8543-6c9722020aba, please check neutron logs for more information. [ 553.520992] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Traceback (most recent call last): [ 553.520992] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 553.520992] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] yield resources [ 553.520992] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 553.520992] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] self.driver.spawn(context, instance, image_meta, [ 553.520992] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 553.520992] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] self._vmops.spawn(context, instance, image_meta, injected_files, [ 553.520992] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 553.520992] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] vm_ref = self.build_virtual_machine(instance, [ 553.521406] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 553.521406] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] vif_infos = vmwarevif.get_vif_info(self._session, [ 553.521406] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 553.521406] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] for vif in network_info: [ 553.521406] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 553.521406] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] return self._sync_wrapper(fn, *args, **kwargs) [ 553.521406] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 553.521406] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] self.wait() [ 553.521406] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 553.521406] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] self[:] = self._gt.wait() [ 553.521406] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 553.521406] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] return self._exit_event.wait() [ 553.521406] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 553.521812] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] result = hub.switch() [ 553.521812] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 553.521812] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] return self.greenlet.switch() [ 553.521812] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.521812] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] result = function(*args, **kwargs) [ 553.521812] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 553.521812] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] return func(*args, **kwargs) [ 553.521812] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 553.521812] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] raise e [ 553.521812] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.521812] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] nwinfo = self.network_api.allocate_for_instance( [ 553.521812] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.521812] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] created_port_ids = self._update_ports_for_instance( [ 553.522193] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.522193] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] with excutils.save_and_reraise_exception(): [ 553.522193] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.522193] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] self.force_reraise() [ 553.522193] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.522193] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] raise self.value [ 553.522193] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.522193] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] updated_port = self._update_port( [ 553.522193] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.522193] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] _ensure_no_port_binding_failure(port) [ 553.522193] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.522193] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] raise exception.PortBindingFailed(port_id=port['id']) [ 553.523835] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] nova.exception.PortBindingFailed: Binding failed for port 7f66cc82-1538-49a5-8543-6c9722020aba, please check neutron logs for more information. [ 553.523835] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] [ 553.523835] env[61867]: INFO nova.compute.manager [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Terminating instance [ 553.523835] env[61867]: DEBUG oslo_concurrency.lockutils [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Acquiring lock "refresh_cache-629da2e8-2f09-42b7-b031-6d8f5a282e37" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.523835] env[61867]: DEBUG oslo_concurrency.lockutils [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Acquired lock "refresh_cache-629da2e8-2f09-42b7-b031-6d8f5a282e37" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.523835] env[61867]: DEBUG nova.network.neutron [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 553.530206] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a61eb1-badf-4fa4-a88b-ae7e1c57cfff {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.540021] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a58f8288-5f32-451b-83cb-f581da3ee00c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.579357] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca60d859-f20e-47e4-8471-ae91b2682838 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.587463] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b07d827d-b5f9-47f8-977c-9a6cf92a0cc1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.605111] env[61867]: DEBUG nova.compute.provider_tree [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 553.796639] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Releasing lock "refresh_cache-433ce856-b04b-4808-8f05-201ad28e18d0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.796920] env[61867]: DEBUG nova.compute.manager [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 553.797362] env[61867]: DEBUG nova.compute.manager [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 553.797570] env[61867]: DEBUG nova.network.neutron [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 553.823643] env[61867]: DEBUG nova.network.neutron [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.879736] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Releasing lock "refresh_cache-c3f0399f-543c-4ab7-8854-b5f7f012fdd2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.881565] env[61867]: DEBUG nova.compute.manager [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 553.881565] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 553.881565] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ecf097b6-d60a-458b-9ce0-aba402837e5e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.894362] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0abcbf77-717b-4c33-88da-df5dcb12f9e2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.923369] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c3f0399f-543c-4ab7-8854-b5f7f012fdd2 could not be found. [ 553.923728] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 553.924282] env[61867]: INFO nova.compute.manager [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 553.924663] env[61867]: DEBUG oslo.service.loopingcall [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 553.925020] env[61867]: DEBUG nova.compute.manager [-] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 553.925212] env[61867]: DEBUG nova.network.neutron [-] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 553.958457] env[61867]: DEBUG nova.network.neutron [-] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.065492] env[61867]: DEBUG nova.network.neutron [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.107882] env[61867]: DEBUG nova.scheduler.client.report [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 554.333599] env[61867]: DEBUG nova.network.neutron [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.380194] env[61867]: DEBUG nova.network.neutron [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.460546] env[61867]: DEBUG nova.network.neutron [-] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.614141] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.614141] env[61867]: DEBUG nova.compute.manager [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 554.616633] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.932s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.838544] env[61867]: INFO nova.compute.manager [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] [instance: 433ce856-b04b-4808-8f05-201ad28e18d0] Took 1.04 seconds to deallocate network for instance. [ 554.884063] env[61867]: DEBUG oslo_concurrency.lockutils [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Releasing lock "refresh_cache-629da2e8-2f09-42b7-b031-6d8f5a282e37" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.885311] env[61867]: DEBUG nova.compute.manager [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 554.885598] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f9a2568-a559-4d84-913d-987c59e1f3aa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.899659] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eba1b48-98cf-4c5b-a57e-fdfe8f3c6d4d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.930599] env[61867]: WARNING nova.virt.vmwareapi.driver [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 629da2e8-2f09-42b7-b031-6d8f5a282e37 could not be found. [ 554.930830] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 554.931136] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b846dd30-440d-4e9f-a460-73e2459d8695 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.939345] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a66806e-cd4a-415f-b456-b20b3afd3b0e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.965816] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 629da2e8-2f09-42b7-b031-6d8f5a282e37 could not be found. [ 554.965816] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 554.965816] env[61867]: INFO nova.compute.manager [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Took 0.08 seconds to destroy the instance on the hypervisor. [ 554.965816] env[61867]: DEBUG oslo.service.loopingcall [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 554.965816] env[61867]: DEBUG nova.compute.manager [-] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 554.965816] env[61867]: DEBUG nova.network.neutron [-] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 554.968538] env[61867]: INFO nova.compute.manager [-] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Took 1.04 seconds to deallocate network for instance. [ 554.974334] env[61867]: DEBUG nova.compute.claims [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 554.974334] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.008145] env[61867]: DEBUG nova.network.neutron [-] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.127219] env[61867]: DEBUG nova.compute.utils [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 555.128579] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Acquiring lock "736f861f-d9a5-4fa5-a28d-26604ebbf051" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.128799] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Lock "736f861f-d9a5-4fa5-a28d-26604ebbf051" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.135248] env[61867]: DEBUG nova.compute.manager [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 555.135248] env[61867]: DEBUG nova.network.neutron [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 555.251165] env[61867]: DEBUG oslo_concurrency.lockutils [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Acquiring lock "d25df58d-4b66-4aa4-9d07-205f36e61775" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.251165] env[61867]: DEBUG oslo_concurrency.lockutils [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Lock "d25df58d-4b66-4aa4-9d07-205f36e61775" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.383373] env[61867]: DEBUG nova.compute.manager [req-21cbd640-5ed5-4b6e-b15e-ea1e6230d031 req-567919d5-eae8-4645-80a7-d2c3510529ae service nova] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Received event network-vif-deleted-770f79ea-89a4-44e5-ad5f-c7fcfb9ffa5b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 555.383373] env[61867]: DEBUG nova.compute.manager [req-21cbd640-5ed5-4b6e-b15e-ea1e6230d031 req-567919d5-eae8-4645-80a7-d2c3510529ae service nova] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Received event network-changed-7f66cc82-1538-49a5-8543-6c9722020aba {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 555.383373] env[61867]: DEBUG nova.compute.manager [req-21cbd640-5ed5-4b6e-b15e-ea1e6230d031 req-567919d5-eae8-4645-80a7-d2c3510529ae service nova] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Refreshing instance network info cache due to event network-changed-7f66cc82-1538-49a5-8543-6c9722020aba. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 555.383373] env[61867]: DEBUG oslo_concurrency.lockutils [req-21cbd640-5ed5-4b6e-b15e-ea1e6230d031 req-567919d5-eae8-4645-80a7-d2c3510529ae service nova] Acquiring lock "refresh_cache-629da2e8-2f09-42b7-b031-6d8f5a282e37" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.383373] env[61867]: DEBUG oslo_concurrency.lockutils [req-21cbd640-5ed5-4b6e-b15e-ea1e6230d031 req-567919d5-eae8-4645-80a7-d2c3510529ae service nova] Acquired lock "refresh_cache-629da2e8-2f09-42b7-b031-6d8f5a282e37" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.383778] env[61867]: DEBUG nova.network.neutron [req-21cbd640-5ed5-4b6e-b15e-ea1e6230d031 req-567919d5-eae8-4645-80a7-d2c3510529ae service nova] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Refreshing network info cache for port 7f66cc82-1538-49a5-8543-6c9722020aba {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 555.476715] env[61867]: DEBUG nova.policy [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '33431e4aba444f90a96022f99cb10a86', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '41b3b859f8d5405fb54b7eb1ccfe6110', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 555.511492] env[61867]: DEBUG nova.network.neutron [-] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.522029] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc37578-abdb-4b44-8980-9948b2a55f6a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.538027] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd572b11-2adb-4ff3-b933-b7cbc34c04f8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.575794] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2448481e-b882-4ff4-9952-1c51cce7d718 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.587111] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1aa50f4-dbf2-4c31-9206-b95a86b1acc8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.603019] env[61867]: DEBUG nova.compute.provider_tree [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 555.636749] env[61867]: DEBUG nova.compute.manager [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 555.899459] env[61867]: INFO nova.scheduler.client.report [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Deleted allocations for instance 433ce856-b04b-4808-8f05-201ad28e18d0 [ 555.945216] env[61867]: DEBUG nova.network.neutron [req-21cbd640-5ed5-4b6e-b15e-ea1e6230d031 req-567919d5-eae8-4645-80a7-d2c3510529ae service nova] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.015580] env[61867]: INFO nova.compute.manager [-] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Took 1.05 seconds to deallocate network for instance. [ 556.109049] env[61867]: DEBUG nova.scheduler.client.report [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 556.250487] env[61867]: DEBUG nova.network.neutron [req-21cbd640-5ed5-4b6e-b15e-ea1e6230d031 req-567919d5-eae8-4645-80a7-d2c3510529ae service nova] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.414427] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b71d4ef-2c81-4f0c-b592-51a1e0aa9224 tempest-ImagesOneServerNegativeTestJSON-1052471862 tempest-ImagesOneServerNegativeTestJSON-1052471862-project-member] Lock "433ce856-b04b-4808-8f05-201ad28e18d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.460s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.543746] env[61867]: ERROR nova.compute.manager [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 59fcbb3a-4369-427e-ab62-f65f5db2eb86, please check neutron logs for more information. [ 556.543746] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 556.543746] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.543746] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 556.543746] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.543746] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 556.543746] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.543746] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 556.543746] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.543746] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 556.543746] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.543746] env[61867]: ERROR nova.compute.manager raise self.value [ 556.543746] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.543746] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 556.543746] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.543746] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 556.544329] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.544329] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 556.544329] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 59fcbb3a-4369-427e-ab62-f65f5db2eb86, please check neutron logs for more information. [ 556.544329] env[61867]: ERROR nova.compute.manager [ 556.544329] env[61867]: Traceback (most recent call last): [ 556.544329] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 556.544329] env[61867]: listener.cb(fileno) [ 556.544329] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.544329] env[61867]: result = function(*args, **kwargs) [ 556.544329] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.544329] env[61867]: return func(*args, **kwargs) [ 556.544329] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.544329] env[61867]: raise e [ 556.544329] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.544329] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 556.544329] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.544329] env[61867]: created_port_ids = self._update_ports_for_instance( [ 556.544329] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.544329] env[61867]: with excutils.save_and_reraise_exception(): [ 556.544329] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.544329] env[61867]: self.force_reraise() [ 556.544329] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.544329] env[61867]: raise self.value [ 556.544329] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.544329] env[61867]: updated_port = self._update_port( [ 556.544329] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.544329] env[61867]: _ensure_no_port_binding_failure(port) [ 556.544329] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.544329] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 556.545196] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 59fcbb3a-4369-427e-ab62-f65f5db2eb86, please check neutron logs for more information. [ 556.545196] env[61867]: Removing descriptor: 21 [ 556.545196] env[61867]: ERROR nova.compute.manager [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 59fcbb3a-4369-427e-ab62-f65f5db2eb86, please check neutron logs for more information. [ 556.545196] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Traceback (most recent call last): [ 556.545196] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 556.545196] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] yield resources [ 556.545196] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 556.545196] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] self.driver.spawn(context, instance, image_meta, [ 556.545196] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 556.545196] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.545196] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.545196] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] vm_ref = self.build_virtual_machine(instance, [ 556.545619] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.545619] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.545619] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.545619] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] for vif in network_info: [ 556.545619] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 556.545619] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] return self._sync_wrapper(fn, *args, **kwargs) [ 556.545619] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 556.545619] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] self.wait() [ 556.545619] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 556.545619] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] self[:] = self._gt.wait() [ 556.545619] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.545619] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] return self._exit_event.wait() [ 556.545619] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.546154] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] result = hub.switch() [ 556.546154] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.546154] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] return self.greenlet.switch() [ 556.546154] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.546154] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] result = function(*args, **kwargs) [ 556.546154] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.546154] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] return func(*args, **kwargs) [ 556.546154] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.546154] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] raise e [ 556.546154] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.546154] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] nwinfo = self.network_api.allocate_for_instance( [ 556.546154] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.546154] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] created_port_ids = self._update_ports_for_instance( [ 556.546659] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.546659] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] with excutils.save_and_reraise_exception(): [ 556.546659] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.546659] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] self.force_reraise() [ 556.546659] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.546659] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] raise self.value [ 556.546659] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.546659] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] updated_port = self._update_port( [ 556.546659] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.546659] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] _ensure_no_port_binding_failure(port) [ 556.546659] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.546659] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] raise exception.PortBindingFailed(port_id=port['id']) [ 556.547119] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] nova.exception.PortBindingFailed: Binding failed for port 59fcbb3a-4369-427e-ab62-f65f5db2eb86, please check neutron logs for more information. [ 556.547119] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] [ 556.547119] env[61867]: INFO nova.compute.manager [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Terminating instance [ 556.547279] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Acquiring lock "refresh_cache-10bd5056-d5b5-48d7-a5b5-bd07dd489618" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.547381] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Acquired lock "refresh_cache-10bd5056-d5b5-48d7-a5b5-bd07dd489618" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.547553] env[61867]: DEBUG nova.network.neutron [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 556.612617] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.996s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.613301] env[61867]: ERROR nova.compute.manager [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b3b1de3c-67f3-4a32-a601-90f3081ce672, please check neutron logs for more information. [ 556.613301] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Traceback (most recent call last): [ 556.613301] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 556.613301] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] self.driver.spawn(context, instance, image_meta, [ 556.613301] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 556.613301] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.613301] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.613301] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] vm_ref = self.build_virtual_machine(instance, [ 556.613301] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.613301] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.613301] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.614477] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] for vif in network_info: [ 556.614477] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 556.614477] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] return self._sync_wrapper(fn, *args, **kwargs) [ 556.614477] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 556.614477] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] self.wait() [ 556.614477] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 556.614477] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] self[:] = self._gt.wait() [ 556.614477] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.614477] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] return self._exit_event.wait() [ 556.614477] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.614477] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] result = hub.switch() [ 556.614477] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.614477] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] return self.greenlet.switch() [ 556.615107] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.615107] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] result = function(*args, **kwargs) [ 556.615107] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.615107] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] return func(*args, **kwargs) [ 556.615107] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.615107] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] raise e [ 556.615107] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.615107] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] nwinfo = self.network_api.allocate_for_instance( [ 556.615107] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.615107] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] created_port_ids = self._update_ports_for_instance( [ 556.615107] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.615107] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] with excutils.save_and_reraise_exception(): [ 556.615107] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.615617] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] self.force_reraise() [ 556.615617] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.615617] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] raise self.value [ 556.615617] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.615617] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] updated_port = self._update_port( [ 556.615617] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.615617] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] _ensure_no_port_binding_failure(port) [ 556.615617] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.615617] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] raise exception.PortBindingFailed(port_id=port['id']) [ 556.615617] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] nova.exception.PortBindingFailed: Binding failed for port b3b1de3c-67f3-4a32-a601-90f3081ce672, please check neutron logs for more information. [ 556.615617] env[61867]: ERROR nova.compute.manager [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] [ 556.616178] env[61867]: DEBUG nova.compute.utils [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Binding failed for port b3b1de3c-67f3-4a32-a601-90f3081ce672, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 556.616178] env[61867]: DEBUG oslo_concurrency.lockutils [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.442s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.616178] env[61867]: DEBUG nova.objects.instance [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61867) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 556.618750] env[61867]: DEBUG nova.compute.manager [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Build of instance ac5df149-83db-481a-a240-e5cf08fdbff9 was re-scheduled: Binding failed for port b3b1de3c-67f3-4a32-a601-90f3081ce672, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 556.618826] env[61867]: DEBUG nova.compute.manager [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 556.619966] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Acquiring lock "refresh_cache-ac5df149-83db-481a-a240-e5cf08fdbff9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.619966] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Acquired lock "refresh_cache-ac5df149-83db-481a-a240-e5cf08fdbff9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.619966] env[61867]: DEBUG nova.network.neutron [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 556.621119] env[61867]: INFO nova.compute.manager [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Took 0.60 seconds to detach 1 volumes for instance. [ 556.623932] env[61867]: DEBUG nova.compute.claims [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 556.623932] env[61867]: DEBUG oslo_concurrency.lockutils [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.649604] env[61867]: DEBUG nova.compute.manager [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 556.689360] env[61867]: DEBUG nova.virt.hardware [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 556.689360] env[61867]: DEBUG nova.virt.hardware [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 556.689360] env[61867]: DEBUG nova.virt.hardware [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 556.689640] env[61867]: DEBUG nova.virt.hardware [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 556.689640] env[61867]: DEBUG nova.virt.hardware [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 556.689640] env[61867]: DEBUG nova.virt.hardware [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 556.692355] env[61867]: DEBUG nova.virt.hardware [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 556.692509] env[61867]: DEBUG nova.virt.hardware [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 556.693159] env[61867]: DEBUG nova.virt.hardware [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 556.693159] env[61867]: DEBUG nova.virt.hardware [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 556.693159] env[61867]: DEBUG nova.virt.hardware [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 556.694334] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62074244-ff69-4b13-80f1-e5905a9ce4b8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.703133] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1727f3f3-0bf8-4c53-bd17-0d0cc7f32341 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.751091] env[61867]: DEBUG nova.network.neutron [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Successfully created port: 697ef278-25e9-4fb5-b582-7dae4b18c7ef {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 556.755482] env[61867]: DEBUG oslo_concurrency.lockutils [req-21cbd640-5ed5-4b6e-b15e-ea1e6230d031 req-567919d5-eae8-4645-80a7-d2c3510529ae service nova] Releasing lock "refresh_cache-629da2e8-2f09-42b7-b031-6d8f5a282e37" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.919417] env[61867]: DEBUG nova.compute.manager [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 557.178713] env[61867]: DEBUG nova.network.neutron [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.273229] env[61867]: DEBUG nova.network.neutron [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.362921] env[61867]: DEBUG nova.network.neutron [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.445429] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.505087] env[61867]: DEBUG nova.network.neutron [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.509428] env[61867]: DEBUG nova.compute.manager [None req-7b34143f-0028-4e31-9a6a-40c8ed91fe8c tempest-ServerDiagnosticsV248Test-1182926949 tempest-ServerDiagnosticsV248Test-1182926949-project-admin] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 557.510738] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-889f2a5c-e317-4dae-bd24-cc404d41be05 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.519026] env[61867]: INFO nova.compute.manager [None req-7b34143f-0028-4e31-9a6a-40c8ed91fe8c tempest-ServerDiagnosticsV248Test-1182926949 tempest-ServerDiagnosticsV248Test-1182926949-project-admin] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Retrieving diagnostics [ 557.521162] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a4f815-8a49-4089-ac72-8eade7402d23 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.636752] env[61867]: DEBUG oslo_concurrency.lockutils [None req-96345111-5d3f-47a5-ba09-b3bc79fbbfc9 tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.021s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.637995] env[61867]: DEBUG oslo_concurrency.lockutils [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.420s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.640574] env[61867]: INFO nova.compute.claims [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 557.866359] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Releasing lock "refresh_cache-ac5df149-83db-481a-a240-e5cf08fdbff9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.866655] env[61867]: DEBUG nova.compute.manager [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 557.866890] env[61867]: DEBUG nova.compute.manager [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.867035] env[61867]: DEBUG nova.network.neutron [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 557.903582] env[61867]: DEBUG nova.network.neutron [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.010725] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Releasing lock "refresh_cache-10bd5056-d5b5-48d7-a5b5-bd07dd489618" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.011040] env[61867]: DEBUG nova.compute.manager [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 558.011040] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 558.011482] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd853bb9-a328-42e3-b7c9-804c1d15c2ca {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.028383] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87a297ee-42a3-4ecb-ab40-b3cf94a25236 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.060647] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 10bd5056-d5b5-48d7-a5b5-bd07dd489618 could not be found. [ 558.060798] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 558.061099] env[61867]: INFO nova.compute.manager [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Took 0.05 seconds to destroy the instance on the hypervisor. [ 558.061194] env[61867]: DEBUG oslo.service.loopingcall [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 558.061781] env[61867]: DEBUG nova.compute.manager [-] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 558.061781] env[61867]: DEBUG nova.network.neutron [-] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 558.092860] env[61867]: DEBUG nova.network.neutron [-] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.101796] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Acquiring lock "ba4cc401-7ae3-40bc-996c-13741a65d991" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.102567] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Lock "ba4cc401-7ae3-40bc-996c-13741a65d991" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.408022] env[61867]: DEBUG nova.network.neutron [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.597772] env[61867]: DEBUG nova.network.neutron [-] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.915367] env[61867]: INFO nova.compute.manager [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] [instance: ac5df149-83db-481a-a240-e5cf08fdbff9] Took 1.04 seconds to deallocate network for instance. [ 559.060635] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a42560e-450b-4888-89d0-4237abf351ae {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.074136] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60ff98e3-8776-4742-87fe-88d7290efacf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.117548] env[61867]: INFO nova.compute.manager [-] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Took 1.06 seconds to deallocate network for instance. [ 559.120404] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b73f1d3-088a-4646-a8f2-ffbf2b880785 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.123515] env[61867]: DEBUG nova.compute.claims [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 559.123515] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.128955] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ace2866-f416-433e-a782-c0f46dba02b3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.144934] env[61867]: DEBUG nova.compute.provider_tree [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 559.174995] env[61867]: DEBUG oslo_concurrency.lockutils [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Acquiring lock "013fc4b2-95f7-439e-b82b-da2265ccf736" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.174995] env[61867]: DEBUG oslo_concurrency.lockutils [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Lock "013fc4b2-95f7-439e-b82b-da2265ccf736" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.398359] env[61867]: DEBUG nova.compute.manager [req-55e8dd9e-94f5-4518-845c-c577fce0c14b req-085e583a-2e8f-461e-9cd0-bdeefd89080b service nova] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Received event network-vif-deleted-7f66cc82-1538-49a5-8543-6c9722020aba {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 559.398359] env[61867]: DEBUG nova.compute.manager [req-55e8dd9e-94f5-4518-845c-c577fce0c14b req-085e583a-2e8f-461e-9cd0-bdeefd89080b service nova] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Received event network-changed-59fcbb3a-4369-427e-ab62-f65f5db2eb86 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 559.398359] env[61867]: DEBUG nova.compute.manager [req-55e8dd9e-94f5-4518-845c-c577fce0c14b req-085e583a-2e8f-461e-9cd0-bdeefd89080b service nova] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Refreshing instance network info cache due to event network-changed-59fcbb3a-4369-427e-ab62-f65f5db2eb86. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 559.398359] env[61867]: DEBUG oslo_concurrency.lockutils [req-55e8dd9e-94f5-4518-845c-c577fce0c14b req-085e583a-2e8f-461e-9cd0-bdeefd89080b service nova] Acquiring lock "refresh_cache-10bd5056-d5b5-48d7-a5b5-bd07dd489618" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.398359] env[61867]: DEBUG oslo_concurrency.lockutils [req-55e8dd9e-94f5-4518-845c-c577fce0c14b req-085e583a-2e8f-461e-9cd0-bdeefd89080b service nova] Acquired lock "refresh_cache-10bd5056-d5b5-48d7-a5b5-bd07dd489618" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.399403] env[61867]: DEBUG nova.network.neutron [req-55e8dd9e-94f5-4518-845c-c577fce0c14b req-085e583a-2e8f-461e-9cd0-bdeefd89080b service nova] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Refreshing network info cache for port 59fcbb3a-4369-427e-ab62-f65f5db2eb86 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 559.537899] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Acquiring lock "32e947ce-c7e5-4568-b1e4-89dfc6a8009e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.538208] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Lock "32e947ce-c7e5-4568-b1e4-89dfc6a8009e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.648920] env[61867]: DEBUG nova.scheduler.client.report [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 559.943725] env[61867]: ERROR nova.compute.manager [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 697ef278-25e9-4fb5-b582-7dae4b18c7ef, please check neutron logs for more information. [ 559.943725] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 559.943725] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.943725] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 559.943725] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 559.943725] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 559.943725] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 559.943725] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 559.943725] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.943725] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 559.943725] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.943725] env[61867]: ERROR nova.compute.manager raise self.value [ 559.943725] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 559.943725] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 559.943725] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.943725] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 559.944420] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.944420] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 559.944420] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 697ef278-25e9-4fb5-b582-7dae4b18c7ef, please check neutron logs for more information. [ 559.944420] env[61867]: ERROR nova.compute.manager [ 559.944420] env[61867]: Traceback (most recent call last): [ 559.944420] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 559.944420] env[61867]: listener.cb(fileno) [ 559.944420] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.944420] env[61867]: result = function(*args, **kwargs) [ 559.944420] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.944420] env[61867]: return func(*args, **kwargs) [ 559.944420] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.944420] env[61867]: raise e [ 559.944420] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.944420] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 559.944420] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 559.944420] env[61867]: created_port_ids = self._update_ports_for_instance( [ 559.944420] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 559.944420] env[61867]: with excutils.save_and_reraise_exception(): [ 559.944420] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.944420] env[61867]: self.force_reraise() [ 559.944420] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.944420] env[61867]: raise self.value [ 559.944420] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 559.944420] env[61867]: updated_port = self._update_port( [ 559.944420] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.944420] env[61867]: _ensure_no_port_binding_failure(port) [ 559.944420] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.944420] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 559.945335] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 697ef278-25e9-4fb5-b582-7dae4b18c7ef, please check neutron logs for more information. [ 559.945335] env[61867]: Removing descriptor: 17 [ 559.945335] env[61867]: ERROR nova.compute.manager [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 697ef278-25e9-4fb5-b582-7dae4b18c7ef, please check neutron logs for more information. [ 559.945335] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Traceback (most recent call last): [ 559.945335] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 559.945335] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] yield resources [ 559.945335] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 559.945335] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] self.driver.spawn(context, instance, image_meta, [ 559.945335] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 559.945335] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 559.945335] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 559.945335] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] vm_ref = self.build_virtual_machine(instance, [ 559.946557] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 559.946557] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] vif_infos = vmwarevif.get_vif_info(self._session, [ 559.946557] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 559.946557] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] for vif in network_info: [ 559.946557] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 559.946557] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] return self._sync_wrapper(fn, *args, **kwargs) [ 559.946557] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 559.946557] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] self.wait() [ 559.946557] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 559.946557] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] self[:] = self._gt.wait() [ 559.946557] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 559.946557] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] return self._exit_event.wait() [ 559.946557] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 559.947099] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] result = hub.switch() [ 559.947099] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 559.947099] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] return self.greenlet.switch() [ 559.947099] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.947099] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] result = function(*args, **kwargs) [ 559.947099] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.947099] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] return func(*args, **kwargs) [ 559.947099] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.947099] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] raise e [ 559.947099] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.947099] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] nwinfo = self.network_api.allocate_for_instance( [ 559.947099] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 559.947099] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] created_port_ids = self._update_ports_for_instance( [ 559.947493] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 559.947493] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] with excutils.save_and_reraise_exception(): [ 559.947493] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.947493] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] self.force_reraise() [ 559.947493] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.947493] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] raise self.value [ 559.947493] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 559.947493] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] updated_port = self._update_port( [ 559.947493] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.947493] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] _ensure_no_port_binding_failure(port) [ 559.947493] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.947493] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] raise exception.PortBindingFailed(port_id=port['id']) [ 559.948230] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] nova.exception.PortBindingFailed: Binding failed for port 697ef278-25e9-4fb5-b582-7dae4b18c7ef, please check neutron logs for more information. [ 559.948230] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] [ 559.948230] env[61867]: INFO nova.compute.manager [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Terminating instance [ 559.951069] env[61867]: DEBUG nova.network.neutron [req-55e8dd9e-94f5-4518-845c-c577fce0c14b req-085e583a-2e8f-461e-9cd0-bdeefd89080b service nova] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.955738] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Acquiring lock "refresh_cache-11292ffa-0db5-4d70-a3c1-31b81bf4182e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.955935] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Acquired lock "refresh_cache-11292ffa-0db5-4d70-a3c1-31b81bf4182e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.956105] env[61867]: DEBUG nova.network.neutron [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 559.977661] env[61867]: INFO nova.scheduler.client.report [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Deleted allocations for instance ac5df149-83db-481a-a240-e5cf08fdbff9 [ 560.115862] env[61867]: DEBUG nova.network.neutron [req-55e8dd9e-94f5-4518-845c-c577fce0c14b req-085e583a-2e8f-461e-9cd0-bdeefd89080b service nova] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.155148] env[61867]: DEBUG oslo_concurrency.lockutils [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.517s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.156167] env[61867]: DEBUG nova.compute.manager [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 560.160067] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.174s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.172844] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Acquiring lock "76d3a228-e314-42b1-9eec-43a32fbd8035" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.173373] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Lock "76d3a228-e314-42b1-9eec-43a32fbd8035" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.173844] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Acquiring lock "76d3a228-e314-42b1-9eec-43a32fbd8035-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.175207] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Lock "76d3a228-e314-42b1-9eec-43a32fbd8035-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.175207] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Lock "76d3a228-e314-42b1-9eec-43a32fbd8035-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.177171] env[61867]: INFO nova.compute.manager [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Terminating instance [ 560.179796] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Acquiring lock "refresh_cache-76d3a228-e314-42b1-9eec-43a32fbd8035" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.179796] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Acquired lock "refresh_cache-76d3a228-e314-42b1-9eec-43a32fbd8035" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.179796] env[61867]: DEBUG nova.network.neutron [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 560.485958] env[61867]: DEBUG nova.network.neutron [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.488414] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9aada1a-f0a8-4b86-a8fa-52f21f0854d6 tempest-ImagesOneServerTestJSON-2030758669 tempest-ImagesOneServerTestJSON-2030758669-project-member] Lock "ac5df149-83db-481a-a240-e5cf08fdbff9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.996s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.619072] env[61867]: DEBUG oslo_concurrency.lockutils [req-55e8dd9e-94f5-4518-845c-c577fce0c14b req-085e583a-2e8f-461e-9cd0-bdeefd89080b service nova] Releasing lock "refresh_cache-10bd5056-d5b5-48d7-a5b5-bd07dd489618" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.620217] env[61867]: DEBUG nova.compute.manager [req-55e8dd9e-94f5-4518-845c-c577fce0c14b req-085e583a-2e8f-461e-9cd0-bdeefd89080b service nova] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Received event network-vif-deleted-59fcbb3a-4369-427e-ab62-f65f5db2eb86 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 560.633711] env[61867]: DEBUG nova.network.neutron [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.667017] env[61867]: DEBUG nova.compute.utils [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 560.671401] env[61867]: DEBUG nova.compute.manager [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 560.671579] env[61867]: DEBUG nova.network.neutron [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 560.712279] env[61867]: DEBUG nova.network.neutron [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.790763] env[61867]: DEBUG nova.policy [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a9666d38328a4f5f9bed0e37895bba66', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0bcfc4ba31f84216beb37a035c3181db', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 560.861826] env[61867]: DEBUG nova.network.neutron [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.986575] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82835faf-bc15-4fb3-a0a2-ac4797027fcf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.990935] env[61867]: DEBUG nova.compute.manager [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 561.000267] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58fd7842-1dfc-4312-8b75-ce90e7407725 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.036850] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d045bb-0730-4ead-a9e7-aebf42a37d79 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.045169] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899b2b0a-befe-4cb6-ba72-a6827cdec661 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.059768] env[61867]: DEBUG nova.compute.provider_tree [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 561.139808] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Releasing lock "refresh_cache-11292ffa-0db5-4d70-a3c1-31b81bf4182e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.139808] env[61867]: DEBUG nova.compute.manager [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 561.139808] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 561.139808] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-df777de6-c66c-472a-b0c3-9c8bdb66aaa5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.150529] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f792ab44-20c1-4192-8f08-c81b9d98f880 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.172928] env[61867]: DEBUG nova.compute.manager [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 561.181663] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 11292ffa-0db5-4d70-a3c1-31b81bf4182e could not be found. [ 561.182354] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 561.182354] env[61867]: INFO nova.compute.manager [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 561.182354] env[61867]: DEBUG oslo.service.loopingcall [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 561.182896] env[61867]: DEBUG nova.compute.manager [-] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 561.182962] env[61867]: DEBUG nova.network.neutron [-] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 561.215077] env[61867]: DEBUG nova.network.neutron [-] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.368676] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Releasing lock "refresh_cache-76d3a228-e314-42b1-9eec-43a32fbd8035" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.369141] env[61867]: DEBUG nova.compute.manager [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 561.369339] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 561.370278] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f145b0c-dbd1-4269-afde-6917e2978d91 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.379505] env[61867]: DEBUG nova.network.neutron [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Successfully created port: b9b17023-6b11-4ffe-9df6-23d94c67b75b {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 561.384576] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 561.385244] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4071d191-265f-4f16-a4c5-4c318ff349cb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.393916] env[61867]: DEBUG oslo_vmware.api [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Waiting for the task: (returnval){ [ 561.393916] env[61867]: value = "task-1276331" [ 561.393916] env[61867]: _type = "Task" [ 561.393916] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.404293] env[61867]: DEBUG oslo_vmware.api [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276331, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.516986] env[61867]: DEBUG oslo_concurrency.lockutils [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.565470] env[61867]: DEBUG nova.scheduler.client.report [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 561.716909] env[61867]: DEBUG nova.network.neutron [-] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.757918] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Acquiring lock "9ffe4fa1-4239-435b-ac7e-648cec92af65" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.758182] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Lock "9ffe4fa1-4239-435b-ac7e-648cec92af65" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.910033] env[61867]: DEBUG oslo_vmware.api [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276331, 'name': PowerOffVM_Task, 'duration_secs': 0.10772} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.910033] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 561.910033] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 561.910033] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc8de5f2-09be-4c95-a6bf-9b6f12b95e93 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.933824] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 561.934171] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 561.934355] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Deleting the datastore file [datastore1] 76d3a228-e314-42b1-9eec-43a32fbd8035 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 561.934985] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4e0328a-8757-4de8-a9fe-90eb394c17c1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.943240] env[61867]: DEBUG oslo_vmware.api [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Waiting for the task: (returnval){ [ 561.943240] env[61867]: value = "task-1276333" [ 561.943240] env[61867]: _type = "Task" [ 561.943240] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.953565] env[61867]: DEBUG oslo_vmware.api [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276333, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.017263] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquiring lock "c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.017495] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.070247] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.910s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.071056] env[61867]: ERROR nova.compute.manager [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b2e60b2d-00eb-419a-b4c7-7c83e016cd65, please check neutron logs for more information. [ 562.071056] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Traceback (most recent call last): [ 562.071056] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 562.071056] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] self.driver.spawn(context, instance, image_meta, [ 562.071056] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 562.071056] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 562.071056] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 562.071056] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] vm_ref = self.build_virtual_machine(instance, [ 562.071056] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 562.071056] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] vif_infos = vmwarevif.get_vif_info(self._session, [ 562.071056] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 562.071603] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] for vif in network_info: [ 562.071603] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 562.071603] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] return self._sync_wrapper(fn, *args, **kwargs) [ 562.071603] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 562.071603] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] self.wait() [ 562.071603] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 562.071603] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] self[:] = self._gt.wait() [ 562.071603] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 562.071603] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] return self._exit_event.wait() [ 562.071603] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 562.071603] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] result = hub.switch() [ 562.071603] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 562.071603] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] return self.greenlet.switch() [ 562.072196] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.072196] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] result = function(*args, **kwargs) [ 562.072196] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 562.072196] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] return func(*args, **kwargs) [ 562.072196] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 562.072196] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] raise e [ 562.072196] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 562.072196] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] nwinfo = self.network_api.allocate_for_instance( [ 562.072196] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 562.072196] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] created_port_ids = self._update_ports_for_instance( [ 562.072196] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 562.072196] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] with excutils.save_and_reraise_exception(): [ 562.072196] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.072817] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] self.force_reraise() [ 562.072817] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.072817] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] raise self.value [ 562.072817] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 562.072817] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] updated_port = self._update_port( [ 562.072817] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.072817] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] _ensure_no_port_binding_failure(port) [ 562.072817] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.072817] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] raise exception.PortBindingFailed(port_id=port['id']) [ 562.072817] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] nova.exception.PortBindingFailed: Binding failed for port b2e60b2d-00eb-419a-b4c7-7c83e016cd65, please check neutron logs for more information. [ 562.072817] env[61867]: ERROR nova.compute.manager [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] [ 562.073142] env[61867]: DEBUG nova.compute.utils [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Binding failed for port b2e60b2d-00eb-419a-b4c7-7c83e016cd65, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 562.073142] env[61867]: DEBUG oslo_concurrency.lockutils [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.923s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.075413] env[61867]: INFO nova.compute.claims [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 562.077827] env[61867]: DEBUG nova.compute.manager [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Build of instance a746fcff-646e-4f9b-b595-7379c6f97efd was re-scheduled: Binding failed for port b2e60b2d-00eb-419a-b4c7-7c83e016cd65, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 562.078424] env[61867]: DEBUG nova.compute.manager [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 562.078501] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "refresh_cache-a746fcff-646e-4f9b-b595-7379c6f97efd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.078591] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquired lock "refresh_cache-a746fcff-646e-4f9b-b595-7379c6f97efd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.078745] env[61867]: DEBUG nova.network.neutron [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 562.191350] env[61867]: DEBUG nova.compute.manager [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 562.224894] env[61867]: DEBUG nova.virt.hardware [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 562.224894] env[61867]: DEBUG nova.virt.hardware [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 562.224894] env[61867]: DEBUG nova.virt.hardware [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 562.225595] env[61867]: DEBUG nova.virt.hardware [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 562.225677] env[61867]: DEBUG nova.virt.hardware [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 562.225798] env[61867]: DEBUG nova.virt.hardware [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 562.226332] env[61867]: DEBUG nova.virt.hardware [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 562.226332] env[61867]: DEBUG nova.virt.hardware [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 562.226332] env[61867]: DEBUG nova.virt.hardware [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 562.226496] env[61867]: DEBUG nova.virt.hardware [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 562.226654] env[61867]: DEBUG nova.virt.hardware [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 562.227267] env[61867]: INFO nova.compute.manager [-] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Took 1.04 seconds to deallocate network for instance. [ 562.227896] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5845af7d-fb02-4d3e-8072-b1997a5e1f72 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.233956] env[61867]: DEBUG nova.compute.claims [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 562.234016] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.238418] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a042c4-7506-4223-a230-d6205df2061b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.458611] env[61867]: DEBUG oslo_vmware.api [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Task: {'id': task-1276333, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095165} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.458611] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 562.458611] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 562.458611] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 562.458611] env[61867]: INFO nova.compute.manager [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Took 1.09 seconds to destroy the instance on the hypervisor. [ 562.458891] env[61867]: DEBUG oslo.service.loopingcall [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 562.458891] env[61867]: DEBUG nova.compute.manager [-] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 562.458891] env[61867]: DEBUG nova.network.neutron [-] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 562.501658] env[61867]: DEBUG nova.network.neutron [-] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.616642] env[61867]: DEBUG nova.network.neutron [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.864435] env[61867]: DEBUG nova.network.neutron [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.012433] env[61867]: DEBUG nova.network.neutron [-] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.080614] env[61867]: DEBUG nova.compute.manager [req-e8b87071-ce7a-4a9e-b04f-3ee361fb1705 req-211000e6-7300-47b9-94ef-509869ade384 service nova] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Received event network-changed-697ef278-25e9-4fb5-b582-7dae4b18c7ef {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 563.080860] env[61867]: DEBUG nova.compute.manager [req-e8b87071-ce7a-4a9e-b04f-3ee361fb1705 req-211000e6-7300-47b9-94ef-509869ade384 service nova] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Refreshing instance network info cache due to event network-changed-697ef278-25e9-4fb5-b582-7dae4b18c7ef. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 563.081553] env[61867]: DEBUG oslo_concurrency.lockutils [req-e8b87071-ce7a-4a9e-b04f-3ee361fb1705 req-211000e6-7300-47b9-94ef-509869ade384 service nova] Acquiring lock "refresh_cache-11292ffa-0db5-4d70-a3c1-31b81bf4182e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.081553] env[61867]: DEBUG oslo_concurrency.lockutils [req-e8b87071-ce7a-4a9e-b04f-3ee361fb1705 req-211000e6-7300-47b9-94ef-509869ade384 service nova] Acquired lock "refresh_cache-11292ffa-0db5-4d70-a3c1-31b81bf4182e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.081553] env[61867]: DEBUG nova.network.neutron [req-e8b87071-ce7a-4a9e-b04f-3ee361fb1705 req-211000e6-7300-47b9-94ef-509869ade384 service nova] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Refreshing network info cache for port 697ef278-25e9-4fb5-b582-7dae4b18c7ef {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 563.369722] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Releasing lock "refresh_cache-a746fcff-646e-4f9b-b595-7379c6f97efd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.369722] env[61867]: DEBUG nova.compute.manager [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 563.369722] env[61867]: DEBUG nova.compute.manager [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 563.369838] env[61867]: DEBUG nova.network.neutron [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 563.432270] env[61867]: DEBUG nova.network.neutron [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.508793] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594db643-05f3-433c-9ee4-03d22b361cf8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.516373] env[61867]: INFO nova.compute.manager [-] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Took 1.06 seconds to deallocate network for instance. [ 563.518986] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19e6b642-df2a-48be-836d-e04c9cebaf09 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.560386] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1507d0-a114-4040-9cef-4edb106bb16b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.570376] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa174a54-0bd1-40a2-9377-447b11f18c52 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.590778] env[61867]: DEBUG nova.compute.provider_tree [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 563.645566] env[61867]: DEBUG nova.network.neutron [req-e8b87071-ce7a-4a9e-b04f-3ee361fb1705 req-211000e6-7300-47b9-94ef-509869ade384 service nova] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.683523] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Acquiring lock "8f8ea2ec-31fe-4726-8f84-89251c138d44" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.683523] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Lock "8f8ea2ec-31fe-4726-8f84-89251c138d44" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.941028] env[61867]: DEBUG nova.network.neutron [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.027238] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.076462] env[61867]: DEBUG nova.network.neutron [req-e8b87071-ce7a-4a9e-b04f-3ee361fb1705 req-211000e6-7300-47b9-94ef-509869ade384 service nova] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.093922] env[61867]: DEBUG nova.scheduler.client.report [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 564.447282] env[61867]: INFO nova.compute.manager [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: a746fcff-646e-4f9b-b595-7379c6f97efd] Took 1.07 seconds to deallocate network for instance. [ 564.521453] env[61867]: ERROR nova.compute.manager [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b9b17023-6b11-4ffe-9df6-23d94c67b75b, please check neutron logs for more information. [ 564.521453] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 564.521453] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 564.521453] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 564.521453] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 564.521453] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 564.521453] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 564.521453] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 564.521453] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.521453] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 564.521453] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.521453] env[61867]: ERROR nova.compute.manager raise self.value [ 564.521453] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 564.521453] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 564.521453] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.521453] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 564.521914] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.521914] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 564.521914] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b9b17023-6b11-4ffe-9df6-23d94c67b75b, please check neutron logs for more information. [ 564.521914] env[61867]: ERROR nova.compute.manager [ 564.521914] env[61867]: Traceback (most recent call last): [ 564.521914] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 564.521914] env[61867]: listener.cb(fileno) [ 564.521914] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 564.521914] env[61867]: result = function(*args, **kwargs) [ 564.521914] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 564.521914] env[61867]: return func(*args, **kwargs) [ 564.521914] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 564.521914] env[61867]: raise e [ 564.521914] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 564.521914] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 564.521914] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 564.521914] env[61867]: created_port_ids = self._update_ports_for_instance( [ 564.521914] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 564.521914] env[61867]: with excutils.save_and_reraise_exception(): [ 564.521914] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.521914] env[61867]: self.force_reraise() [ 564.521914] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.521914] env[61867]: raise self.value [ 564.521914] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 564.521914] env[61867]: updated_port = self._update_port( [ 564.521914] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.521914] env[61867]: _ensure_no_port_binding_failure(port) [ 564.521914] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.521914] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 564.522705] env[61867]: nova.exception.PortBindingFailed: Binding failed for port b9b17023-6b11-4ffe-9df6-23d94c67b75b, please check neutron logs for more information. [ 564.522705] env[61867]: Removing descriptor: 21 [ 564.522705] env[61867]: ERROR nova.compute.manager [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b9b17023-6b11-4ffe-9df6-23d94c67b75b, please check neutron logs for more information. [ 564.522705] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Traceback (most recent call last): [ 564.522705] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 564.522705] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] yield resources [ 564.522705] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 564.522705] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] self.driver.spawn(context, instance, image_meta, [ 564.522705] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 564.522705] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] self._vmops.spawn(context, instance, image_meta, injected_files, [ 564.522705] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 564.522705] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] vm_ref = self.build_virtual_machine(instance, [ 564.523254] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 564.523254] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] vif_infos = vmwarevif.get_vif_info(self._session, [ 564.523254] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 564.523254] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] for vif in network_info: [ 564.523254] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 564.523254] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] return self._sync_wrapper(fn, *args, **kwargs) [ 564.523254] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 564.523254] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] self.wait() [ 564.523254] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 564.523254] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] self[:] = self._gt.wait() [ 564.523254] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 564.523254] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] return self._exit_event.wait() [ 564.523254] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 564.523619] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] result = hub.switch() [ 564.523619] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 564.523619] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] return self.greenlet.switch() [ 564.523619] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 564.523619] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] result = function(*args, **kwargs) [ 564.523619] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 564.523619] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] return func(*args, **kwargs) [ 564.523619] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 564.523619] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] raise e [ 564.523619] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 564.523619] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] nwinfo = self.network_api.allocate_for_instance( [ 564.523619] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 564.523619] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] created_port_ids = self._update_ports_for_instance( [ 564.523997] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 564.523997] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] with excutils.save_and_reraise_exception(): [ 564.523997] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.523997] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] self.force_reraise() [ 564.523997] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.523997] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] raise self.value [ 564.523997] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 564.523997] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] updated_port = self._update_port( [ 564.523997] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.523997] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] _ensure_no_port_binding_failure(port) [ 564.523997] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.523997] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] raise exception.PortBindingFailed(port_id=port['id']) [ 564.524357] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] nova.exception.PortBindingFailed: Binding failed for port b9b17023-6b11-4ffe-9df6-23d94c67b75b, please check neutron logs for more information. [ 564.524357] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] [ 564.524357] env[61867]: INFO nova.compute.manager [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Terminating instance [ 564.525394] env[61867]: DEBUG oslo_concurrency.lockutils [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Acquiring lock "refresh_cache-f911465c-277a-41ee-a6b5-0e3b85185b56" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 564.525394] env[61867]: DEBUG oslo_concurrency.lockutils [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Acquired lock "refresh_cache-f911465c-277a-41ee-a6b5-0e3b85185b56" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 564.525715] env[61867]: DEBUG nova.network.neutron [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 564.580214] env[61867]: DEBUG oslo_concurrency.lockutils [req-e8b87071-ce7a-4a9e-b04f-3ee361fb1705 req-211000e6-7300-47b9-94ef-509869ade384 service nova] Releasing lock "refresh_cache-11292ffa-0db5-4d70-a3c1-31b81bf4182e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.580214] env[61867]: DEBUG nova.compute.manager [req-e8b87071-ce7a-4a9e-b04f-3ee361fb1705 req-211000e6-7300-47b9-94ef-509869ade384 service nova] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Received event network-vif-deleted-697ef278-25e9-4fb5-b582-7dae4b18c7ef {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 564.606641] env[61867]: DEBUG oslo_concurrency.lockutils [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.534s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.607312] env[61867]: DEBUG nova.compute.manager [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 564.612238] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.415s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.612238] env[61867]: DEBUG nova.objects.instance [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61867) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 565.067055] env[61867]: DEBUG nova.network.neutron [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.121595] env[61867]: DEBUG nova.compute.utils [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 565.127407] env[61867]: DEBUG nova.compute.manager [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 565.128348] env[61867]: DEBUG nova.network.neutron [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 565.341743] env[61867]: DEBUG nova.policy [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ea8a14b0ae374c4cad41618fd9c8ad91', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd49c191097244596b0f10ec9b53c95ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 565.351579] env[61867]: DEBUG nova.network.neutron [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.484022] env[61867]: INFO nova.scheduler.client.report [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Deleted allocations for instance a746fcff-646e-4f9b-b595-7379c6f97efd [ 565.499169] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Acquiring lock "50930c54-d03e-4529-ac85-a73e9a5b4e4d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.499740] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Lock "50930c54-d03e-4529-ac85-a73e9a5b4e4d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.631664] env[61867]: DEBUG nova.compute.manager [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 565.636816] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a713e87f-828f-441b-b519-fc06d0704364 tempest-ServersAdmin275Test-83054379 tempest-ServersAdmin275Test-83054379-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.025s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.640032] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.584s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.640032] env[61867]: INFO nova.compute.claims [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 565.855576] env[61867]: DEBUG oslo_concurrency.lockutils [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Releasing lock "refresh_cache-f911465c-277a-41ee-a6b5-0e3b85185b56" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.856017] env[61867]: DEBUG nova.compute.manager [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 565.856308] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 565.859044] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-657779c3-beaa-4849-83ff-62ae9856fbfa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.862217] env[61867]: DEBUG nova.network.neutron [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Successfully created port: 425c84f1-a9a7-4602-b422-a01b089bab3e {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 565.871177] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42da897b-93c9-4998-aaf1-8932d2d60255 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.894334] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Acquiring lock "572dbc25-39a6-4e8b-b5c0-176ea0e27839" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.894574] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Lock "572dbc25-39a6-4e8b-b5c0-176ea0e27839" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.900613] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f911465c-277a-41ee-a6b5-0e3b85185b56 could not be found. [ 565.900613] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 565.900613] env[61867]: INFO nova.compute.manager [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Took 0.04 seconds to destroy the instance on the hypervisor. [ 565.900810] env[61867]: DEBUG oslo.service.loopingcall [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 565.900859] env[61867]: DEBUG nova.compute.manager [-] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 565.900970] env[61867]: DEBUG nova.network.neutron [-] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 565.929269] env[61867]: DEBUG nova.network.neutron [-] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.995269] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bbebfce8-5458-4f50-ba35-dd048e0f738d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "a746fcff-646e-4f9b-b595-7379c6f97efd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.718s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.432787] env[61867]: DEBUG nova.network.neutron [-] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.499911] env[61867]: DEBUG nova.compute.manager [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 566.647636] env[61867]: DEBUG nova.compute.manager [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 566.682083] env[61867]: DEBUG nova.virt.hardware [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 566.682083] env[61867]: DEBUG nova.virt.hardware [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 566.682083] env[61867]: DEBUG nova.virt.hardware [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 566.682333] env[61867]: DEBUG nova.virt.hardware [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 566.682333] env[61867]: DEBUG nova.virt.hardware [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 566.682333] env[61867]: DEBUG nova.virt.hardware [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 566.682574] env[61867]: DEBUG nova.virt.hardware [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 566.682574] env[61867]: DEBUG nova.virt.hardware [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 566.682677] env[61867]: DEBUG nova.virt.hardware [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 566.684287] env[61867]: DEBUG nova.virt.hardware [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 566.684287] env[61867]: DEBUG nova.virt.hardware [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 566.685748] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73debc2c-aa6a-471f-9397-86f415dcfc1c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.697777] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3160ef1-cdbb-4aed-bb90-9fbcfecec103 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.782248] env[61867]: DEBUG nova.compute.manager [req-8d80306e-f20a-4f3d-afb3-e7bd9a5769d9 req-98eac9c7-effe-4b5e-989e-dde7e76c47fd service nova] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Received event network-changed-b9b17023-6b11-4ffe-9df6-23d94c67b75b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 566.782441] env[61867]: DEBUG nova.compute.manager [req-8d80306e-f20a-4f3d-afb3-e7bd9a5769d9 req-98eac9c7-effe-4b5e-989e-dde7e76c47fd service nova] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Refreshing instance network info cache due to event network-changed-b9b17023-6b11-4ffe-9df6-23d94c67b75b. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 566.782649] env[61867]: DEBUG oslo_concurrency.lockutils [req-8d80306e-f20a-4f3d-afb3-e7bd9a5769d9 req-98eac9c7-effe-4b5e-989e-dde7e76c47fd service nova] Acquiring lock "refresh_cache-f911465c-277a-41ee-a6b5-0e3b85185b56" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.782794] env[61867]: DEBUG oslo_concurrency.lockutils [req-8d80306e-f20a-4f3d-afb3-e7bd9a5769d9 req-98eac9c7-effe-4b5e-989e-dde7e76c47fd service nova] Acquired lock "refresh_cache-f911465c-277a-41ee-a6b5-0e3b85185b56" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.782990] env[61867]: DEBUG nova.network.neutron [req-8d80306e-f20a-4f3d-afb3-e7bd9a5769d9 req-98eac9c7-effe-4b5e-989e-dde7e76c47fd service nova] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Refreshing network info cache for port b9b17023-6b11-4ffe-9df6-23d94c67b75b {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 566.936613] env[61867]: INFO nova.compute.manager [-] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Took 1.04 seconds to deallocate network for instance. [ 566.944291] env[61867]: DEBUG nova.compute.claims [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 566.944291] env[61867]: DEBUG oslo_concurrency.lockutils [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.026573] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.098682] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "274c5625-eb81-45d1-bd95-7336bddfad1e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.100130] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "274c5625-eb81-45d1-bd95-7336bddfad1e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.106197] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f95d49ba-3269-4f50-8456-f42f14ab0f98 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.116985] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a4d79d9-4cb0-462e-a2d0-9a590d18c5a9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.154019] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c502c00e-50df-44fa-8368-f7ef3464ee09 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.162982] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31594d64-f44e-4d13-bd23-994f05e464c9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.177321] env[61867]: DEBUG nova.compute.provider_tree [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 567.318113] env[61867]: DEBUG nova.network.neutron [req-8d80306e-f20a-4f3d-afb3-e7bd9a5769d9 req-98eac9c7-effe-4b5e-989e-dde7e76c47fd service nova] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.680512] env[61867]: DEBUG nova.scheduler.client.report [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 567.747916] env[61867]: DEBUG nova.network.neutron [req-8d80306e-f20a-4f3d-afb3-e7bd9a5769d9 req-98eac9c7-effe-4b5e-989e-dde7e76c47fd service nova] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.190152] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.552s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.190777] env[61867]: DEBUG nova.compute.manager [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 568.194024] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.926s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.194378] env[61867]: DEBUG nova.objects.instance [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Lazy-loading 'resources' on Instance uuid 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 568.253701] env[61867]: DEBUG oslo_concurrency.lockutils [req-8d80306e-f20a-4f3d-afb3-e7bd9a5769d9 req-98eac9c7-effe-4b5e-989e-dde7e76c47fd service nova] Releasing lock "refresh_cache-f911465c-277a-41ee-a6b5-0e3b85185b56" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.253701] env[61867]: DEBUG nova.compute.manager [req-8d80306e-f20a-4f3d-afb3-e7bd9a5769d9 req-98eac9c7-effe-4b5e-989e-dde7e76c47fd service nova] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Received event network-vif-deleted-b9b17023-6b11-4ffe-9df6-23d94c67b75b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 568.705868] env[61867]: DEBUG nova.compute.utils [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 568.712779] env[61867]: DEBUG nova.compute.manager [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 568.712779] env[61867]: DEBUG nova.network.neutron [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 568.938068] env[61867]: DEBUG nova.policy [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'be73da428e26490bae2ae5e6f6d64cb8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c48448efedee4a81be27f2128b1b5f2a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 569.142660] env[61867]: DEBUG nova.compute.manager [req-96df2111-3f85-4ffd-a91c-bdfeb2a490ee req-b32ddf19-72ac-47bb-83f8-fc7cfcaa7ffb service nova] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Received event network-changed-425c84f1-a9a7-4602-b422-a01b089bab3e {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 569.142868] env[61867]: DEBUG nova.compute.manager [req-96df2111-3f85-4ffd-a91c-bdfeb2a490ee req-b32ddf19-72ac-47bb-83f8-fc7cfcaa7ffb service nova] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Refreshing instance network info cache due to event network-changed-425c84f1-a9a7-4602-b422-a01b089bab3e. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 569.143172] env[61867]: DEBUG oslo_concurrency.lockutils [req-96df2111-3f85-4ffd-a91c-bdfeb2a490ee req-b32ddf19-72ac-47bb-83f8-fc7cfcaa7ffb service nova] Acquiring lock "refresh_cache-814dd6d2-43b1-4700-b585-c5d33b96931f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.144124] env[61867]: DEBUG oslo_concurrency.lockutils [req-96df2111-3f85-4ffd-a91c-bdfeb2a490ee req-b32ddf19-72ac-47bb-83f8-fc7cfcaa7ffb service nova] Acquired lock "refresh_cache-814dd6d2-43b1-4700-b585-c5d33b96931f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.144124] env[61867]: DEBUG nova.network.neutron [req-96df2111-3f85-4ffd-a91c-bdfeb2a490ee req-b32ddf19-72ac-47bb-83f8-fc7cfcaa7ffb service nova] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Refreshing network info cache for port 425c84f1-a9a7-4602-b422-a01b089bab3e {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 569.219508] env[61867]: DEBUG nova.compute.manager [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 569.304822] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5bf589-09e6-4f6e-98f3-dcd5d20429bc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.315124] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25311d50-054a-4efb-bc5b-cfc3472e95af {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.357243] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa73657-c524-4121-85f7-21d44c9d1eff {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.365513] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b1cab06-5056-456d-8c21-b675677a9899 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.384176] env[61867]: DEBUG nova.compute.provider_tree [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.421643] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "aad1e5ea-5a9c-4499-8f67-2516c5c285cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.421643] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "aad1e5ea-5a9c-4499-8f67-2516c5c285cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.569262] env[61867]: ERROR nova.compute.manager [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 425c84f1-a9a7-4602-b422-a01b089bab3e, please check neutron logs for more information. [ 569.569262] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 569.569262] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.569262] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 569.569262] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 569.569262] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 569.569262] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 569.569262] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 569.569262] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.569262] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 569.569262] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.569262] env[61867]: ERROR nova.compute.manager raise self.value [ 569.569262] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 569.569262] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 569.569262] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.569262] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 569.569806] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.569806] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 569.569806] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 425c84f1-a9a7-4602-b422-a01b089bab3e, please check neutron logs for more information. [ 569.569806] env[61867]: ERROR nova.compute.manager [ 569.569806] env[61867]: Traceback (most recent call last): [ 569.569806] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 569.569806] env[61867]: listener.cb(fileno) [ 569.569806] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.569806] env[61867]: result = function(*args, **kwargs) [ 569.569806] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 569.569806] env[61867]: return func(*args, **kwargs) [ 569.569806] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.569806] env[61867]: raise e [ 569.569806] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.569806] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 569.569806] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 569.569806] env[61867]: created_port_ids = self._update_ports_for_instance( [ 569.569806] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 569.569806] env[61867]: with excutils.save_and_reraise_exception(): [ 569.569806] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.569806] env[61867]: self.force_reraise() [ 569.569806] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.569806] env[61867]: raise self.value [ 569.569806] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 569.569806] env[61867]: updated_port = self._update_port( [ 569.569806] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.569806] env[61867]: _ensure_no_port_binding_failure(port) [ 569.569806] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.569806] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 569.570758] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 425c84f1-a9a7-4602-b422-a01b089bab3e, please check neutron logs for more information. [ 569.570758] env[61867]: Removing descriptor: 17 [ 569.570758] env[61867]: ERROR nova.compute.manager [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 425c84f1-a9a7-4602-b422-a01b089bab3e, please check neutron logs for more information. [ 569.570758] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Traceback (most recent call last): [ 569.570758] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 569.570758] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] yield resources [ 569.570758] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 569.570758] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] self.driver.spawn(context, instance, image_meta, [ 569.570758] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 569.570758] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 569.570758] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 569.570758] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] vm_ref = self.build_virtual_machine(instance, [ 569.571471] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 569.571471] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] vif_infos = vmwarevif.get_vif_info(self._session, [ 569.571471] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 569.571471] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] for vif in network_info: [ 569.571471] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 569.571471] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] return self._sync_wrapper(fn, *args, **kwargs) [ 569.571471] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 569.571471] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] self.wait() [ 569.571471] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 569.571471] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] self[:] = self._gt.wait() [ 569.571471] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 569.571471] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] return self._exit_event.wait() [ 569.571471] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 569.571973] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] result = hub.switch() [ 569.571973] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 569.571973] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] return self.greenlet.switch() [ 569.571973] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.571973] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] result = function(*args, **kwargs) [ 569.571973] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 569.571973] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] return func(*args, **kwargs) [ 569.571973] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.571973] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] raise e [ 569.571973] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.571973] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] nwinfo = self.network_api.allocate_for_instance( [ 569.571973] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 569.571973] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] created_port_ids = self._update_ports_for_instance( [ 569.572764] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 569.572764] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] with excutils.save_and_reraise_exception(): [ 569.572764] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.572764] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] self.force_reraise() [ 569.572764] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.572764] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] raise self.value [ 569.572764] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 569.572764] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] updated_port = self._update_port( [ 569.572764] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.572764] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] _ensure_no_port_binding_failure(port) [ 569.572764] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.572764] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] raise exception.PortBindingFailed(port_id=port['id']) [ 569.573180] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] nova.exception.PortBindingFailed: Binding failed for port 425c84f1-a9a7-4602-b422-a01b089bab3e, please check neutron logs for more information. [ 569.573180] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] [ 569.573180] env[61867]: INFO nova.compute.manager [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Terminating instance [ 569.573180] env[61867]: DEBUG oslo_concurrency.lockutils [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Acquiring lock "refresh_cache-814dd6d2-43b1-4700-b585-c5d33b96931f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.691228] env[61867]: DEBUG nova.network.neutron [req-96df2111-3f85-4ffd-a91c-bdfeb2a490ee req-b32ddf19-72ac-47bb-83f8-fc7cfcaa7ffb service nova] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 569.887466] env[61867]: DEBUG nova.scheduler.client.report [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 569.912237] env[61867]: DEBUG nova.network.neutron [req-96df2111-3f85-4ffd-a91c-bdfeb2a490ee req-b32ddf19-72ac-47bb-83f8-fc7cfcaa7ffb service nova] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.202589] env[61867]: DEBUG nova.network.neutron [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Successfully created port: baaff768-f304-4adb-a80f-87f99daa6bfd {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 570.234596] env[61867]: DEBUG nova.compute.manager [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 570.275924] env[61867]: DEBUG nova.virt.hardware [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 570.275924] env[61867]: DEBUG nova.virt.hardware [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 570.275924] env[61867]: DEBUG nova.virt.hardware [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 570.275924] env[61867]: DEBUG nova.virt.hardware [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 570.276185] env[61867]: DEBUG nova.virt.hardware [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 570.276185] env[61867]: DEBUG nova.virt.hardware [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 570.276185] env[61867]: DEBUG nova.virt.hardware [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 570.276185] env[61867]: DEBUG nova.virt.hardware [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 570.277113] env[61867]: DEBUG nova.virt.hardware [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 570.277373] env[61867]: DEBUG nova.virt.hardware [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 570.277468] env[61867]: DEBUG nova.virt.hardware [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 570.278908] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50283d9-9325-4d32-9bab-280471b52cce {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.291122] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92499f44-73e9-4256-b1d4-378cde5df2b1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.394752] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.199s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.397393] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.425s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.415253] env[61867]: DEBUG oslo_concurrency.lockutils [req-96df2111-3f85-4ffd-a91c-bdfeb2a490ee req-b32ddf19-72ac-47bb-83f8-fc7cfcaa7ffb service nova] Releasing lock "refresh_cache-814dd6d2-43b1-4700-b585-c5d33b96931f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.415653] env[61867]: DEBUG oslo_concurrency.lockutils [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Acquired lock "refresh_cache-814dd6d2-43b1-4700-b585-c5d33b96931f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 570.415837] env[61867]: DEBUG nova.network.neutron [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 570.432042] env[61867]: INFO nova.scheduler.client.report [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Deleted allocations for instance 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa [ 570.738089] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquiring lock "37ed9e18-8dba-459d-bc67-e3ce3f9cdaea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.738089] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Lock "37ed9e18-8dba-459d-bc67-e3ce3f9cdaea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.946518] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dabeff6f-3356-403d-81eb-902da502874e tempest-ServersAdmin275Test-1512287911 tempest-ServersAdmin275Test-1512287911-project-member] Lock "318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.759s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.951622] env[61867]: DEBUG nova.network.neutron [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.039091] env[61867]: DEBUG nova.network.neutron [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.240206] env[61867]: DEBUG nova.compute.manager [req-0a971dec-efd3-4452-8f00-d58491393e7d req-fb8ab06a-72ee-4a2a-a9d8-ec5f41c18aa9 service nova] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Received event network-vif-deleted-425c84f1-a9a7-4602-b422-a01b089bab3e {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 571.415631] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d57e14f-f016-46b3-871d-167367929dc9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.424886] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f9301ae-4e89-4aba-b54c-38e00c3e7c5f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.460606] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7517fe69-d24d-43d8-8416-b0dc5e39cea2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.469369] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad52888-617a-4814-ba02-4e0859f95e71 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.483609] env[61867]: DEBUG nova.compute.provider_tree [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.542800] env[61867]: DEBUG oslo_concurrency.lockutils [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Releasing lock "refresh_cache-814dd6d2-43b1-4700-b585-c5d33b96931f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.542800] env[61867]: DEBUG nova.compute.manager [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 571.542800] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 571.542800] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b61bc33c-4ea1-428a-a54a-5ae590b04a87 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.555853] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6c3701-0548-4d12-81da-55aa137f29bc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.584533] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 814dd6d2-43b1-4700-b585-c5d33b96931f could not be found. [ 571.585029] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 571.585275] env[61867]: INFO nova.compute.manager [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 571.585566] env[61867]: DEBUG oslo.service.loopingcall [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 571.585809] env[61867]: DEBUG nova.compute.manager [-] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 571.585938] env[61867]: DEBUG nova.network.neutron [-] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 571.621619] env[61867]: DEBUG nova.network.neutron [-] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.986360] env[61867]: DEBUG nova.scheduler.client.report [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.125166] env[61867]: DEBUG nova.network.neutron [-] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.142268] env[61867]: ERROR nova.compute.manager [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port baaff768-f304-4adb-a80f-87f99daa6bfd, please check neutron logs for more information. [ 572.142268] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 572.142268] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.142268] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 572.142268] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.142268] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 572.142268] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.142268] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 572.142268] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.142268] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 572.142268] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.142268] env[61867]: ERROR nova.compute.manager raise self.value [ 572.142268] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.142268] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 572.142268] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.142268] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 572.142787] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.142787] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 572.142787] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port baaff768-f304-4adb-a80f-87f99daa6bfd, please check neutron logs for more information. [ 572.142787] env[61867]: ERROR nova.compute.manager [ 572.142787] env[61867]: Traceback (most recent call last): [ 572.142787] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 572.142787] env[61867]: listener.cb(fileno) [ 572.142787] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.142787] env[61867]: result = function(*args, **kwargs) [ 572.142787] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 572.142787] env[61867]: return func(*args, **kwargs) [ 572.142787] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.142787] env[61867]: raise e [ 572.142787] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.142787] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 572.142787] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.142787] env[61867]: created_port_ids = self._update_ports_for_instance( [ 572.142787] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.142787] env[61867]: with excutils.save_and_reraise_exception(): [ 572.142787] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.142787] env[61867]: self.force_reraise() [ 572.142787] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.142787] env[61867]: raise self.value [ 572.142787] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.142787] env[61867]: updated_port = self._update_port( [ 572.142787] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.142787] env[61867]: _ensure_no_port_binding_failure(port) [ 572.142787] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.142787] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 572.143736] env[61867]: nova.exception.PortBindingFailed: Binding failed for port baaff768-f304-4adb-a80f-87f99daa6bfd, please check neutron logs for more information. [ 572.143736] env[61867]: Removing descriptor: 21 [ 572.143736] env[61867]: ERROR nova.compute.manager [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port baaff768-f304-4adb-a80f-87f99daa6bfd, please check neutron logs for more information. [ 572.143736] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Traceback (most recent call last): [ 572.143736] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 572.143736] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] yield resources [ 572.143736] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 572.143736] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] self.driver.spawn(context, instance, image_meta, [ 572.143736] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 572.143736] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.143736] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.143736] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] vm_ref = self.build_virtual_machine(instance, [ 572.144134] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.144134] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.144134] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.144134] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] for vif in network_info: [ 572.144134] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 572.144134] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] return self._sync_wrapper(fn, *args, **kwargs) [ 572.144134] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 572.144134] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] self.wait() [ 572.144134] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 572.144134] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] self[:] = self._gt.wait() [ 572.144134] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.144134] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] return self._exit_event.wait() [ 572.144134] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.144498] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] result = hub.switch() [ 572.144498] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.144498] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] return self.greenlet.switch() [ 572.144498] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.144498] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] result = function(*args, **kwargs) [ 572.144498] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 572.144498] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] return func(*args, **kwargs) [ 572.144498] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.144498] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] raise e [ 572.144498] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.144498] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] nwinfo = self.network_api.allocate_for_instance( [ 572.144498] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.144498] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] created_port_ids = self._update_ports_for_instance( [ 572.144950] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.144950] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] with excutils.save_and_reraise_exception(): [ 572.144950] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.144950] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] self.force_reraise() [ 572.144950] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.144950] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] raise self.value [ 572.144950] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.144950] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] updated_port = self._update_port( [ 572.144950] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.144950] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] _ensure_no_port_binding_failure(port) [ 572.144950] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.144950] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] raise exception.PortBindingFailed(port_id=port['id']) [ 572.145294] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] nova.exception.PortBindingFailed: Binding failed for port baaff768-f304-4adb-a80f-87f99daa6bfd, please check neutron logs for more information. [ 572.145294] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] [ 572.145294] env[61867]: INFO nova.compute.manager [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Terminating instance [ 572.149809] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Acquiring lock "refresh_cache-0d716b01-a5d5-40f6-b002-5e1488bd54af" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.149809] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Acquired lock "refresh_cache-0d716b01-a5d5-40f6-b002-5e1488bd54af" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.149809] env[61867]: DEBUG nova.network.neutron [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 572.492945] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.095s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.493598] env[61867]: ERROR nova.compute.manager [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 770f79ea-89a4-44e5-ad5f-c7fcfb9ffa5b, please check neutron logs for more information. [ 572.493598] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Traceback (most recent call last): [ 572.493598] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 572.493598] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] self.driver.spawn(context, instance, image_meta, [ 572.493598] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 572.493598] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.493598] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.493598] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] vm_ref = self.build_virtual_machine(instance, [ 572.493598] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.493598] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.493598] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.493943] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] for vif in network_info: [ 572.493943] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 572.493943] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] return self._sync_wrapper(fn, *args, **kwargs) [ 572.493943] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 572.493943] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] self.wait() [ 572.493943] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 572.493943] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] self[:] = self._gt.wait() [ 572.493943] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.493943] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] return self._exit_event.wait() [ 572.493943] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.493943] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] result = hub.switch() [ 572.493943] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.493943] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] return self.greenlet.switch() [ 572.494332] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.494332] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] result = function(*args, **kwargs) [ 572.494332] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 572.494332] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] return func(*args, **kwargs) [ 572.494332] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.494332] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] raise e [ 572.494332] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.494332] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] nwinfo = self.network_api.allocate_for_instance( [ 572.494332] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.494332] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] created_port_ids = self._update_ports_for_instance( [ 572.494332] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.494332] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] with excutils.save_and_reraise_exception(): [ 572.494332] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.494714] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] self.force_reraise() [ 572.494714] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.494714] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] raise self.value [ 572.494714] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.494714] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] updated_port = self._update_port( [ 572.494714] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.494714] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] _ensure_no_port_binding_failure(port) [ 572.494714] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.494714] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] raise exception.PortBindingFailed(port_id=port['id']) [ 572.494714] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] nova.exception.PortBindingFailed: Binding failed for port 770f79ea-89a4-44e5-ad5f-c7fcfb9ffa5b, please check neutron logs for more information. [ 572.494714] env[61867]: ERROR nova.compute.manager [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] [ 572.495075] env[61867]: DEBUG nova.compute.utils [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Binding failed for port 770f79ea-89a4-44e5-ad5f-c7fcfb9ffa5b, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 572.496347] env[61867]: DEBUG oslo_concurrency.lockutils [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.873s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.501244] env[61867]: DEBUG nova.compute.manager [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Build of instance c3f0399f-543c-4ab7-8854-b5f7f012fdd2 was re-scheduled: Binding failed for port 770f79ea-89a4-44e5-ad5f-c7fcfb9ffa5b, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 572.501244] env[61867]: DEBUG nova.compute.manager [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 572.501567] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Acquiring lock "refresh_cache-c3f0399f-543c-4ab7-8854-b5f7f012fdd2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.501567] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Acquired lock "refresh_cache-c3f0399f-543c-4ab7-8854-b5f7f012fdd2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.501704] env[61867]: DEBUG nova.network.neutron [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 572.627562] env[61867]: INFO nova.compute.manager [-] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Took 1.04 seconds to deallocate network for instance. [ 572.630294] env[61867]: DEBUG nova.compute.claims [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 572.630294] env[61867]: DEBUG oslo_concurrency.lockutils [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.681827] env[61867]: DEBUG nova.network.neutron [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.119895] env[61867]: DEBUG nova.network.neutron [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.126570] env[61867]: DEBUG nova.network.neutron [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.321972] env[61867]: DEBUG nova.compute.manager [req-5cd37f49-600d-47f3-9e74-d48c7cebecd6 req-1a5bba77-cd7d-47cb-a6e7-2d3eb66695b4 service nova] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Received event network-changed-baaff768-f304-4adb-a80f-87f99daa6bfd {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 573.322612] env[61867]: DEBUG nova.compute.manager [req-5cd37f49-600d-47f3-9e74-d48c7cebecd6 req-1a5bba77-cd7d-47cb-a6e7-2d3eb66695b4 service nova] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Refreshing instance network info cache due to event network-changed-baaff768-f304-4adb-a80f-87f99daa6bfd. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 573.322612] env[61867]: DEBUG oslo_concurrency.lockutils [req-5cd37f49-600d-47f3-9e74-d48c7cebecd6 req-1a5bba77-cd7d-47cb-a6e7-2d3eb66695b4 service nova] Acquiring lock "refresh_cache-0d716b01-a5d5-40f6-b002-5e1488bd54af" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.427982] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-940a8a31-f885-4bae-9c58-1e240614336f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.432909] env[61867]: DEBUG nova.network.neutron [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.443715] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc96fd97-fced-4ea3-acba-f976f9173c06 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.489436] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef29f9de-112c-4eff-a243-98be1bf89519 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.501905] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d6d9f1a-bc10-4668-95b3-5256302cfe98 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.516792] env[61867]: DEBUG nova.compute.provider_tree [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.632323] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Releasing lock "refresh_cache-0d716b01-a5d5-40f6-b002-5e1488bd54af" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.633071] env[61867]: DEBUG nova.compute.manager [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 573.633071] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 573.633317] env[61867]: DEBUG oslo_concurrency.lockutils [req-5cd37f49-600d-47f3-9e74-d48c7cebecd6 req-1a5bba77-cd7d-47cb-a6e7-2d3eb66695b4 service nova] Acquired lock "refresh_cache-0d716b01-a5d5-40f6-b002-5e1488bd54af" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.633568] env[61867]: DEBUG nova.network.neutron [req-5cd37f49-600d-47f3-9e74-d48c7cebecd6 req-1a5bba77-cd7d-47cb-a6e7-2d3eb66695b4 service nova] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Refreshing network info cache for port baaff768-f304-4adb-a80f-87f99daa6bfd {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 573.634830] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-394a6d0f-d4cb-40bc-ad76-bd9455fb1d19 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.645332] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b519ea-bdd5-4a89-bf84-284b303b3a68 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.674755] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0d716b01-a5d5-40f6-b002-5e1488bd54af could not be found. [ 573.675040] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 573.675263] env[61867]: INFO nova.compute.manager [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Took 0.04 seconds to destroy the instance on the hypervisor. [ 573.675520] env[61867]: DEBUG oslo.service.loopingcall [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 573.675771] env[61867]: DEBUG nova.compute.manager [-] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 573.675867] env[61867]: DEBUG nova.network.neutron [-] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 573.701647] env[61867]: DEBUG nova.network.neutron [-] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.939893] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Releasing lock "refresh_cache-c3f0399f-543c-4ab7-8854-b5f7f012fdd2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.940111] env[61867]: DEBUG nova.compute.manager [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 573.941016] env[61867]: DEBUG nova.compute.manager [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 573.941016] env[61867]: DEBUG nova.network.neutron [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 573.976246] env[61867]: DEBUG nova.network.neutron [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.023393] env[61867]: DEBUG nova.scheduler.client.report [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 574.176242] env[61867]: DEBUG nova.network.neutron [req-5cd37f49-600d-47f3-9e74-d48c7cebecd6 req-1a5bba77-cd7d-47cb-a6e7-2d3eb66695b4 service nova] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.203866] env[61867]: DEBUG nova.network.neutron [-] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.319977] env[61867]: DEBUG nova.network.neutron [req-5cd37f49-600d-47f3-9e74-d48c7cebecd6 req-1a5bba77-cd7d-47cb-a6e7-2d3eb66695b4 service nova] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.482714] env[61867]: DEBUG nova.network.neutron [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.530025] env[61867]: DEBUG oslo_concurrency.lockutils [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.033s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.530025] env[61867]: ERROR nova.compute.manager [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7f66cc82-1538-49a5-8543-6c9722020aba, please check neutron logs for more information. [ 574.530025] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Traceback (most recent call last): [ 574.530025] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 574.530025] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] self.driver.spawn(context, instance, image_meta, [ 574.530025] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 574.530025] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] self._vmops.spawn(context, instance, image_meta, injected_files, [ 574.530025] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 574.530025] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] vm_ref = self.build_virtual_machine(instance, [ 574.530414] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 574.530414] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] vif_infos = vmwarevif.get_vif_info(self._session, [ 574.530414] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 574.530414] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] for vif in network_info: [ 574.530414] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 574.530414] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] return self._sync_wrapper(fn, *args, **kwargs) [ 574.530414] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 574.530414] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] self.wait() [ 574.530414] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 574.530414] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] self[:] = self._gt.wait() [ 574.530414] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 574.530414] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] return self._exit_event.wait() [ 574.530414] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 574.530817] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] result = hub.switch() [ 574.530817] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 574.530817] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] return self.greenlet.switch() [ 574.530817] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.530817] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] result = function(*args, **kwargs) [ 574.530817] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 574.530817] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] return func(*args, **kwargs) [ 574.530817] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 574.530817] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] raise e [ 574.530817] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.530817] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] nwinfo = self.network_api.allocate_for_instance( [ 574.530817] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 574.530817] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] created_port_ids = self._update_ports_for_instance( [ 574.531217] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 574.531217] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] with excutils.save_and_reraise_exception(): [ 574.531217] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.531217] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] self.force_reraise() [ 574.531217] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.531217] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] raise self.value [ 574.531217] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 574.531217] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] updated_port = self._update_port( [ 574.531217] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.531217] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] _ensure_no_port_binding_failure(port) [ 574.531217] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.531217] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] raise exception.PortBindingFailed(port_id=port['id']) [ 574.531598] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] nova.exception.PortBindingFailed: Binding failed for port 7f66cc82-1538-49a5-8543-6c9722020aba, please check neutron logs for more information. [ 574.531598] env[61867]: ERROR nova.compute.manager [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] [ 574.533736] env[61867]: DEBUG nova.compute.utils [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Binding failed for port 7f66cc82-1538-49a5-8543-6c9722020aba, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 574.533736] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.089s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.535629] env[61867]: INFO nova.compute.claims [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 574.538979] env[61867]: DEBUG nova.compute.manager [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Build of instance 629da2e8-2f09-42b7-b031-6d8f5a282e37 was re-scheduled: Binding failed for port 7f66cc82-1538-49a5-8543-6c9722020aba, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 574.541215] env[61867]: DEBUG nova.compute.manager [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 574.541215] env[61867]: DEBUG oslo_concurrency.lockutils [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Acquiring lock "refresh_cache-629da2e8-2f09-42b7-b031-6d8f5a282e37" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.541215] env[61867]: DEBUG oslo_concurrency.lockutils [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Acquired lock "refresh_cache-629da2e8-2f09-42b7-b031-6d8f5a282e37" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.541215] env[61867]: DEBUG nova.network.neutron [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 574.712878] env[61867]: INFO nova.compute.manager [-] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Took 1.03 seconds to deallocate network for instance. [ 574.716439] env[61867]: DEBUG nova.compute.claims [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 574.720496] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.801258] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "5c250675-3708-4c25-b370-6707c457357a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.801513] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "5c250675-3708-4c25-b370-6707c457357a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.825222] env[61867]: DEBUG oslo_concurrency.lockutils [req-5cd37f49-600d-47f3-9e74-d48c7cebecd6 req-1a5bba77-cd7d-47cb-a6e7-2d3eb66695b4 service nova] Releasing lock "refresh_cache-0d716b01-a5d5-40f6-b002-5e1488bd54af" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.825222] env[61867]: DEBUG nova.compute.manager [req-5cd37f49-600d-47f3-9e74-d48c7cebecd6 req-1a5bba77-cd7d-47cb-a6e7-2d3eb66695b4 service nova] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Received event network-vif-deleted-baaff768-f304-4adb-a80f-87f99daa6bfd {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 574.985599] env[61867]: INFO nova.compute.manager [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] [instance: c3f0399f-543c-4ab7-8854-b5f7f012fdd2] Took 1.04 seconds to deallocate network for instance. [ 575.070330] env[61867]: DEBUG nova.network.neutron [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.422350] env[61867]: DEBUG nova.network.neutron [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.923713] env[61867]: DEBUG oslo_concurrency.lockutils [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Releasing lock "refresh_cache-629da2e8-2f09-42b7-b031-6d8f5a282e37" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.926901] env[61867]: DEBUG nova.compute.manager [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 575.927326] env[61867]: DEBUG nova.compute.manager [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 575.927668] env[61867]: DEBUG nova.network.neutron [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 575.964529] env[61867]: DEBUG nova.network.neutron [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.029906] env[61867]: DEBUG oslo_concurrency.lockutils [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "4d9c1c85-9d49-4c7c-bcac-f22f5842db0d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.030503] env[61867]: DEBUG oslo_concurrency.lockutils [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "4d9c1c85-9d49-4c7c-bcac-f22f5842db0d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.040768] env[61867]: INFO nova.scheduler.client.report [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Deleted allocations for instance c3f0399f-543c-4ab7-8854-b5f7f012fdd2 [ 576.155199] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304cb01f-7537-4630-9bb9-8a5ce0158692 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.164496] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de073fa-6658-4e74-82e6-af1c480bd3e4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.207303] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dade5cfb-1fa0-4583-a3cb-318cb2085fab {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.216666] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ddbca2-6511-4f12-a2cd-bf33bcb74cbf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.233979] env[61867]: DEBUG nova.compute.provider_tree [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.467488] env[61867]: DEBUG nova.network.neutron [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.554492] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd46c8f7-1c7a-4e7c-8859-8eb19610d251 tempest-ServersWithSpecificFlavorTestJSON-1386586042 tempest-ServersWithSpecificFlavorTestJSON-1386586042-project-member] Lock "c3f0399f-543c-4ab7-8854-b5f7f012fdd2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.930s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.739176] env[61867]: DEBUG nova.scheduler.client.report [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 576.971378] env[61867]: INFO nova.compute.manager [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] [instance: 629da2e8-2f09-42b7-b031-6d8f5a282e37] Took 1.04 seconds to deallocate network for instance. [ 577.063639] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 577.248638] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.715s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.249713] env[61867]: DEBUG nova.compute.manager [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 577.255177] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.131s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.446113] env[61867]: DEBUG oslo_concurrency.lockutils [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Acquiring lock "59d2ee3a-756f-4dda-a70a-6d43aaf32049" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.446387] env[61867]: DEBUG oslo_concurrency.lockutils [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Lock "59d2ee3a-756f-4dda-a70a-6d43aaf32049" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.592481] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.761230] env[61867]: DEBUG nova.compute.utils [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 577.768302] env[61867]: DEBUG nova.compute.manager [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 577.768302] env[61867]: DEBUG nova.network.neutron [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 577.839987] env[61867]: DEBUG nova.policy [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '624eae2c3c6f49e09bee42d57e40be2e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3234e3e4e57549d6a6af972a05fb7448', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 577.907401] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Acquiring lock "774e7c38-426a-441d-8252-3daa424be558" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.907623] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Lock "774e7c38-426a-441d-8252-3daa424be558" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.015459] env[61867]: INFO nova.scheduler.client.report [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Deleted allocations for instance 629da2e8-2f09-42b7-b031-6d8f5a282e37 [ 578.227476] env[61867]: DEBUG nova.network.neutron [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Successfully created port: 007ca2b7-ee1b-4705-94e9-e47938ebe5bf {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 578.274276] env[61867]: DEBUG nova.compute.manager [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 578.299052] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dfeffd6-dc5f-4cca-b868-e8138bad6efc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.307647] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc71518-3a67-4a09-b787-7d09093c2c8e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.341075] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b416adea-fd4d-4af5-b28c-c5529dc47831 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.347900] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6143b1cc-be70-4497-82fe-8efeccebd3fd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.362884] env[61867]: DEBUG nova.compute.provider_tree [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 578.532620] env[61867]: DEBUG oslo_concurrency.lockutils [None req-48a52bcd-fcc2-4869-95a3-a710bf573f93 tempest-ServerActionsV293TestJSON-1852684811 tempest-ServerActionsV293TestJSON-1852684811-project-member] Lock "629da2e8-2f09-42b7-b031-6d8f5a282e37" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.656s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.866630] env[61867]: DEBUG nova.scheduler.client.report [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 579.037023] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 579.286131] env[61867]: DEBUG nova.compute.manager [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 579.313844] env[61867]: DEBUG nova.virt.hardware [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 579.316445] env[61867]: DEBUG nova.virt.hardware [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 579.316445] env[61867]: DEBUG nova.virt.hardware [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 579.316445] env[61867]: DEBUG nova.virt.hardware [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 579.316445] env[61867]: DEBUG nova.virt.hardware [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 579.316445] env[61867]: DEBUG nova.virt.hardware [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 579.316681] env[61867]: DEBUG nova.virt.hardware [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 579.316681] env[61867]: DEBUG nova.virt.hardware [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 579.316681] env[61867]: DEBUG nova.virt.hardware [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 579.316681] env[61867]: DEBUG nova.virt.hardware [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 579.316681] env[61867]: DEBUG nova.virt.hardware [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 579.316990] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a6c0cb9-c978-4cb0-82e1-a70e3dd2d2c1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.326519] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b07955-9e04-47f1-ba40-0f08e8962bd8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.371841] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.117s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.371841] env[61867]: ERROR nova.compute.manager [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 59fcbb3a-4369-427e-ab62-f65f5db2eb86, please check neutron logs for more information. [ 579.371841] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Traceback (most recent call last): [ 579.371841] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 579.371841] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] self.driver.spawn(context, instance, image_meta, [ 579.371841] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 579.371841] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] self._vmops.spawn(context, instance, image_meta, injected_files, [ 579.371841] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 579.371841] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] vm_ref = self.build_virtual_machine(instance, [ 579.372205] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 579.372205] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] vif_infos = vmwarevif.get_vif_info(self._session, [ 579.372205] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 579.372205] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] for vif in network_info: [ 579.372205] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 579.372205] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] return self._sync_wrapper(fn, *args, **kwargs) [ 579.372205] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 579.372205] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] self.wait() [ 579.372205] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 579.372205] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] self[:] = self._gt.wait() [ 579.372205] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 579.372205] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] return self._exit_event.wait() [ 579.372205] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 579.372679] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] result = hub.switch() [ 579.372679] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 579.372679] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] return self.greenlet.switch() [ 579.372679] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.372679] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] result = function(*args, **kwargs) [ 579.372679] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.372679] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] return func(*args, **kwargs) [ 579.372679] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.372679] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] raise e [ 579.372679] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.372679] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] nwinfo = self.network_api.allocate_for_instance( [ 579.372679] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.372679] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] created_port_ids = self._update_ports_for_instance( [ 579.373063] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.373063] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] with excutils.save_and_reraise_exception(): [ 579.373063] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.373063] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] self.force_reraise() [ 579.373063] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.373063] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] raise self.value [ 579.373063] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.373063] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] updated_port = self._update_port( [ 579.373063] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.373063] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] _ensure_no_port_binding_failure(port) [ 579.373063] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.373063] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] raise exception.PortBindingFailed(port_id=port['id']) [ 579.373475] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] nova.exception.PortBindingFailed: Binding failed for port 59fcbb3a-4369-427e-ab62-f65f5db2eb86, please check neutron logs for more information. [ 579.373475] env[61867]: ERROR nova.compute.manager [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] [ 579.373475] env[61867]: DEBUG nova.compute.utils [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Binding failed for port 59fcbb3a-4369-427e-ab62-f65f5db2eb86, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 579.373769] env[61867]: DEBUG oslo_concurrency.lockutils [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.857s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.375314] env[61867]: INFO nova.compute.claims [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 579.377869] env[61867]: DEBUG nova.compute.manager [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Build of instance 10bd5056-d5b5-48d7-a5b5-bd07dd489618 was re-scheduled: Binding failed for port 59fcbb3a-4369-427e-ab62-f65f5db2eb86, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 579.379878] env[61867]: DEBUG nova.compute.manager [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 579.379878] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Acquiring lock "refresh_cache-10bd5056-d5b5-48d7-a5b5-bd07dd489618" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.379878] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Acquired lock "refresh_cache-10bd5056-d5b5-48d7-a5b5-bd07dd489618" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.379878] env[61867]: DEBUG nova.network.neutron [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 579.497227] env[61867]: ERROR nova.compute.manager [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 007ca2b7-ee1b-4705-94e9-e47938ebe5bf, please check neutron logs for more information. [ 579.497227] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 579.497227] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.497227] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 579.497227] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.497227] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 579.497227] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.497227] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 579.497227] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.497227] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 579.497227] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.497227] env[61867]: ERROR nova.compute.manager raise self.value [ 579.497227] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.497227] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 579.497227] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.497227] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 579.497826] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.497826] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 579.497826] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 007ca2b7-ee1b-4705-94e9-e47938ebe5bf, please check neutron logs for more information. [ 579.497826] env[61867]: ERROR nova.compute.manager [ 579.497826] env[61867]: Traceback (most recent call last): [ 579.497826] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 579.497826] env[61867]: listener.cb(fileno) [ 579.497826] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.497826] env[61867]: result = function(*args, **kwargs) [ 579.497826] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.497826] env[61867]: return func(*args, **kwargs) [ 579.497826] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.497826] env[61867]: raise e [ 579.497826] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.497826] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 579.497826] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.497826] env[61867]: created_port_ids = self._update_ports_for_instance( [ 579.497826] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.497826] env[61867]: with excutils.save_and_reraise_exception(): [ 579.497826] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.497826] env[61867]: self.force_reraise() [ 579.497826] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.497826] env[61867]: raise self.value [ 579.497826] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.497826] env[61867]: updated_port = self._update_port( [ 579.497826] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.497826] env[61867]: _ensure_no_port_binding_failure(port) [ 579.497826] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.497826] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 579.498750] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 007ca2b7-ee1b-4705-94e9-e47938ebe5bf, please check neutron logs for more information. [ 579.498750] env[61867]: Removing descriptor: 17 [ 579.499118] env[61867]: ERROR nova.compute.manager [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 007ca2b7-ee1b-4705-94e9-e47938ebe5bf, please check neutron logs for more information. [ 579.499118] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Traceback (most recent call last): [ 579.499118] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 579.499118] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] yield resources [ 579.499118] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 579.499118] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] self.driver.spawn(context, instance, image_meta, [ 579.499118] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 579.499118] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] self._vmops.spawn(context, instance, image_meta, injected_files, [ 579.499118] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 579.499118] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] vm_ref = self.build_virtual_machine(instance, [ 579.499118] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 579.499517] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] vif_infos = vmwarevif.get_vif_info(self._session, [ 579.499517] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 579.499517] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] for vif in network_info: [ 579.499517] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 579.499517] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] return self._sync_wrapper(fn, *args, **kwargs) [ 579.499517] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 579.499517] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] self.wait() [ 579.499517] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 579.499517] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] self[:] = self._gt.wait() [ 579.499517] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 579.499517] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] return self._exit_event.wait() [ 579.499517] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 579.499517] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] result = hub.switch() [ 579.500121] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 579.500121] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] return self.greenlet.switch() [ 579.500121] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.500121] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] result = function(*args, **kwargs) [ 579.500121] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.500121] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] return func(*args, **kwargs) [ 579.500121] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.500121] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] raise e [ 579.500121] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.500121] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] nwinfo = self.network_api.allocate_for_instance( [ 579.500121] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.500121] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] created_port_ids = self._update_ports_for_instance( [ 579.500121] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.501769] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] with excutils.save_and_reraise_exception(): [ 579.501769] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.501769] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] self.force_reraise() [ 579.501769] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.501769] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] raise self.value [ 579.501769] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.501769] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] updated_port = self._update_port( [ 579.501769] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.501769] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] _ensure_no_port_binding_failure(port) [ 579.501769] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.501769] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] raise exception.PortBindingFailed(port_id=port['id']) [ 579.501769] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] nova.exception.PortBindingFailed: Binding failed for port 007ca2b7-ee1b-4705-94e9-e47938ebe5bf, please check neutron logs for more information. [ 579.501769] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] [ 579.502159] env[61867]: INFO nova.compute.manager [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Terminating instance [ 579.503439] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquiring lock "refresh_cache-76c297c0-f393-499a-80e5-a78d84ec0885" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.503647] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquired lock "refresh_cache-76c297c0-f393-499a-80e5-a78d84ec0885" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.504215] env[61867]: DEBUG nova.network.neutron [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 579.560529] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.909806] env[61867]: DEBUG nova.network.neutron [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.011888] env[61867]: DEBUG nova.network.neutron [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.024198] env[61867]: DEBUG nova.network.neutron [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.102408] env[61867]: DEBUG nova.compute.manager [req-b49b9725-8265-44b8-8145-1c540b96adf5 req-d949d5e3-fc5d-4a27-91f8-a63f3ec88ada service nova] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Received event network-changed-007ca2b7-ee1b-4705-94e9-e47938ebe5bf {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 580.102627] env[61867]: DEBUG nova.compute.manager [req-b49b9725-8265-44b8-8145-1c540b96adf5 req-d949d5e3-fc5d-4a27-91f8-a63f3ec88ada service nova] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Refreshing instance network info cache due to event network-changed-007ca2b7-ee1b-4705-94e9-e47938ebe5bf. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 580.102814] env[61867]: DEBUG oslo_concurrency.lockutils [req-b49b9725-8265-44b8-8145-1c540b96adf5 req-d949d5e3-fc5d-4a27-91f8-a63f3ec88ada service nova] Acquiring lock "refresh_cache-76c297c0-f393-499a-80e5-a78d84ec0885" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.104762] env[61867]: DEBUG nova.network.neutron [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.514916] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Releasing lock "refresh_cache-10bd5056-d5b5-48d7-a5b5-bd07dd489618" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.515229] env[61867]: DEBUG nova.compute.manager [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 580.515462] env[61867]: DEBUG nova.compute.manager [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 580.515633] env[61867]: DEBUG nova.network.neutron [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 580.538933] env[61867]: DEBUG nova.network.neutron [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.610136] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Releasing lock "refresh_cache-76c297c0-f393-499a-80e5-a78d84ec0885" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.610462] env[61867]: DEBUG nova.compute.manager [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 580.610655] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 580.613508] env[61867]: DEBUG oslo_concurrency.lockutils [req-b49b9725-8265-44b8-8145-1c540b96adf5 req-d949d5e3-fc5d-4a27-91f8-a63f3ec88ada service nova] Acquired lock "refresh_cache-76c297c0-f393-499a-80e5-a78d84ec0885" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.613911] env[61867]: DEBUG nova.network.neutron [req-b49b9725-8265-44b8-8145-1c540b96adf5 req-d949d5e3-fc5d-4a27-91f8-a63f3ec88ada service nova] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Refreshing network info cache for port 007ca2b7-ee1b-4705-94e9-e47938ebe5bf {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 580.615012] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e526305b-7872-4607-be0d-d601a9c462cc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.624869] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a6a5b6-3260-4cf4-9022-670bcc28dbb1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.652451] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 76c297c0-f393-499a-80e5-a78d84ec0885 could not be found. [ 580.652967] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 580.652967] env[61867]: INFO nova.compute.manager [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Took 0.04 seconds to destroy the instance on the hypervisor. [ 580.653104] env[61867]: DEBUG oslo.service.loopingcall [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 580.653311] env[61867]: DEBUG nova.compute.manager [-] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 580.653406] env[61867]: DEBUG nova.network.neutron [-] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 580.675533] env[61867]: DEBUG nova.network.neutron [-] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.811601] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d72190-0659-4920-bc45-2ab1c21c8899 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.819252] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f889a15a-d7fa-42ba-af59-f3803869070b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.853027] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f626d09-30c5-4a26-b35d-ff16f5dc98a7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.860765] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e1d548-2b8f-4920-8880-22e5d5ad0399 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.878616] env[61867]: DEBUG nova.compute.provider_tree [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.045729] env[61867]: DEBUG nova.network.neutron [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.150247] env[61867]: DEBUG nova.network.neutron [req-b49b9725-8265-44b8-8145-1c540b96adf5 req-d949d5e3-fc5d-4a27-91f8-a63f3ec88ada service nova] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 581.178708] env[61867]: DEBUG nova.network.neutron [-] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.337407] env[61867]: DEBUG nova.network.neutron [req-b49b9725-8265-44b8-8145-1c540b96adf5 req-d949d5e3-fc5d-4a27-91f8-a63f3ec88ada service nova] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.381305] env[61867]: DEBUG nova.scheduler.client.report [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 581.547470] env[61867]: INFO nova.compute.manager [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] [instance: 10bd5056-d5b5-48d7-a5b5-bd07dd489618] Took 1.03 seconds to deallocate network for instance. [ 581.682083] env[61867]: INFO nova.compute.manager [-] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Took 1.03 seconds to deallocate network for instance. [ 581.690945] env[61867]: DEBUG nova.compute.claims [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 581.690945] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.840682] env[61867]: DEBUG oslo_concurrency.lockutils [req-b49b9725-8265-44b8-8145-1c540b96adf5 req-d949d5e3-fc5d-4a27-91f8-a63f3ec88ada service nova] Releasing lock "refresh_cache-76c297c0-f393-499a-80e5-a78d84ec0885" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.886805] env[61867]: DEBUG oslo_concurrency.lockutils [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.887964] env[61867]: DEBUG nova.compute.manager [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 581.890189] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.656s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.401256] env[61867]: DEBUG nova.compute.utils [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 582.405869] env[61867]: DEBUG nova.compute.manager [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 582.406069] env[61867]: DEBUG nova.network.neutron [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 582.586390] env[61867]: INFO nova.scheduler.client.report [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Deleted allocations for instance 10bd5056-d5b5-48d7-a5b5-bd07dd489618 [ 582.703285] env[61867]: DEBUG nova.compute.manager [req-014325af-7d41-48bb-aa8a-5a50c61fca60 req-2625a76f-1773-4bd0-8e93-62fa321c3bb7 service nova] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Received event network-vif-deleted-007ca2b7-ee1b-4705-94e9-e47938ebe5bf {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 582.750118] env[61867]: DEBUG nova.policy [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '624eae2c3c6f49e09bee42d57e40be2e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3234e3e4e57549d6a6af972a05fb7448', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 582.896154] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f19cb3d4-d1e6-43b2-8758-62844617dfac {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.901196] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d87ea09-a90a-4db8-b969-ad4a52ba068f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.933983] env[61867]: DEBUG nova.compute.manager [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 582.938398] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411e0698-8cbb-4be2-84c4-ce7c1fc6564c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.951527] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50dae43-3245-4ca9-bfe4-89b3114fe0fb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.969772] env[61867]: DEBUG nova.compute.provider_tree [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.096587] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cb6ea175-94fd-4b1c-9215-bd2432adc389 tempest-VolumesAssistedSnapshotsTest-521549723 tempest-VolumesAssistedSnapshotsTest-521549723-project-member] Lock "10bd5056-d5b5-48d7-a5b5-bd07dd489618" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.355s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.475019] env[61867]: DEBUG nova.scheduler.client.report [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 583.584095] env[61867]: DEBUG nova.network.neutron [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Successfully created port: 5a8a05c3-4601-48bc-a242-f4627a6e0d1a {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 583.600024] env[61867]: DEBUG nova.compute.manager [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 583.943970] env[61867]: DEBUG nova.compute.manager [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 583.983434] env[61867]: DEBUG nova.virt.hardware [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 583.983987] env[61867]: DEBUG nova.virt.hardware [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 583.984205] env[61867]: DEBUG nova.virt.hardware [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 583.984732] env[61867]: DEBUG nova.virt.hardware [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 583.984732] env[61867]: DEBUG nova.virt.hardware [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 583.984732] env[61867]: DEBUG nova.virt.hardware [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 583.984960] env[61867]: DEBUG nova.virt.hardware [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 583.985075] env[61867]: DEBUG nova.virt.hardware [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 583.985208] env[61867]: DEBUG nova.virt.hardware [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 583.985367] env[61867]: DEBUG nova.virt.hardware [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 583.985536] env[61867]: DEBUG nova.virt.hardware [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 583.986434] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.096s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.988797] env[61867]: ERROR nova.compute.manager [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 697ef278-25e9-4fb5-b582-7dae4b18c7ef, please check neutron logs for more information. [ 583.988797] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Traceback (most recent call last): [ 583.988797] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 583.988797] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] self.driver.spawn(context, instance, image_meta, [ 583.988797] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 583.988797] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 583.988797] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 583.988797] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] vm_ref = self.build_virtual_machine(instance, [ 583.988797] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 583.988797] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] vif_infos = vmwarevif.get_vif_info(self._session, [ 583.988797] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 583.989800] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] for vif in network_info: [ 583.989800] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 583.989800] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] return self._sync_wrapper(fn, *args, **kwargs) [ 583.989800] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 583.989800] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] self.wait() [ 583.989800] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 583.989800] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] self[:] = self._gt.wait() [ 583.989800] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 583.989800] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] return self._exit_event.wait() [ 583.989800] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 583.989800] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] result = hub.switch() [ 583.989800] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 583.989800] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] return self.greenlet.switch() [ 583.990296] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.990296] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] result = function(*args, **kwargs) [ 583.990296] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 583.990296] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] return func(*args, **kwargs) [ 583.990296] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 583.990296] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] raise e [ 583.990296] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.990296] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] nwinfo = self.network_api.allocate_for_instance( [ 583.990296] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.990296] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] created_port_ids = self._update_ports_for_instance( [ 583.990296] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.990296] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] with excutils.save_and_reraise_exception(): [ 583.990296] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.990742] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] self.force_reraise() [ 583.990742] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.990742] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] raise self.value [ 583.990742] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.990742] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] updated_port = self._update_port( [ 583.990742] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.990742] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] _ensure_no_port_binding_failure(port) [ 583.990742] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.990742] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] raise exception.PortBindingFailed(port_id=port['id']) [ 583.990742] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] nova.exception.PortBindingFailed: Binding failed for port 697ef278-25e9-4fb5-b582-7dae4b18c7ef, please check neutron logs for more information. [ 583.990742] env[61867]: ERROR nova.compute.manager [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] [ 583.990962] env[61867]: DEBUG nova.compute.utils [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Binding failed for port 697ef278-25e9-4fb5-b582-7dae4b18c7ef, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 583.990962] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db3719b9-7e53-46a5-8382-cae54765633a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.993619] env[61867]: DEBUG nova.compute.manager [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Build of instance 11292ffa-0db5-4d70-a3c1-31b81bf4182e was re-scheduled: Binding failed for port 697ef278-25e9-4fb5-b582-7dae4b18c7ef, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 583.994098] env[61867]: DEBUG nova.compute.manager [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 583.994328] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Acquiring lock "refresh_cache-11292ffa-0db5-4d70-a3c1-31b81bf4182e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.994477] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Acquired lock "refresh_cache-11292ffa-0db5-4d70-a3c1-31b81bf4182e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.994646] env[61867]: DEBUG nova.network.neutron [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 583.995674] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.969s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.995893] env[61867]: DEBUG nova.objects.instance [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Lazy-loading 'resources' on Instance uuid 76d3a228-e314-42b1-9eec-43a32fbd8035 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 584.005824] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1ed9cf-b573-4991-98f9-a69e1ba4d1ee {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.130884] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.523479] env[61867]: DEBUG nova.network.neutron [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.672960] env[61867]: DEBUG nova.network.neutron [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.989080] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea1727c-9f68-4493-bc54-e23950d82165 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.999317] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b1a47c-fff2-4a58-a369-a307f64eb704 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.041828] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa7bb7e-ea79-4bbd-bc7d-6a5d7fcf904b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.051056] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-418a7152-3e53-4cc1-9edc-8b5479b94232 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.066931] env[61867]: DEBUG nova.compute.provider_tree [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.178871] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Releasing lock "refresh_cache-11292ffa-0db5-4d70-a3c1-31b81bf4182e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.179141] env[61867]: DEBUG nova.compute.manager [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 585.179325] env[61867]: DEBUG nova.compute.manager [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 585.179495] env[61867]: DEBUG nova.network.neutron [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 585.203655] env[61867]: DEBUG nova.network.neutron [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.420984] env[61867]: ERROR nova.compute.manager [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5a8a05c3-4601-48bc-a242-f4627a6e0d1a, please check neutron logs for more information. [ 585.420984] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 585.420984] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.420984] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 585.420984] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.420984] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 585.420984] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.420984] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 585.420984] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.420984] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 585.420984] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.420984] env[61867]: ERROR nova.compute.manager raise self.value [ 585.420984] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.420984] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 585.420984] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.420984] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 585.421390] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.421390] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 585.421390] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5a8a05c3-4601-48bc-a242-f4627a6e0d1a, please check neutron logs for more information. [ 585.421390] env[61867]: ERROR nova.compute.manager [ 585.421390] env[61867]: Traceback (most recent call last): [ 585.421390] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 585.421390] env[61867]: listener.cb(fileno) [ 585.421390] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.421390] env[61867]: result = function(*args, **kwargs) [ 585.421390] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.421390] env[61867]: return func(*args, **kwargs) [ 585.421390] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.421390] env[61867]: raise e [ 585.421390] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.421390] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 585.421390] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.421390] env[61867]: created_port_ids = self._update_ports_for_instance( [ 585.421390] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.421390] env[61867]: with excutils.save_and_reraise_exception(): [ 585.421390] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.421390] env[61867]: self.force_reraise() [ 585.421390] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.421390] env[61867]: raise self.value [ 585.421390] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.421390] env[61867]: updated_port = self._update_port( [ 585.421390] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.421390] env[61867]: _ensure_no_port_binding_failure(port) [ 585.421390] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.421390] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 585.422033] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 5a8a05c3-4601-48bc-a242-f4627a6e0d1a, please check neutron logs for more information. [ 585.422033] env[61867]: Removing descriptor: 21 [ 585.422033] env[61867]: ERROR nova.compute.manager [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5a8a05c3-4601-48bc-a242-f4627a6e0d1a, please check neutron logs for more information. [ 585.422033] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Traceback (most recent call last): [ 585.422033] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 585.422033] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] yield resources [ 585.422033] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 585.422033] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] self.driver.spawn(context, instance, image_meta, [ 585.422033] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 585.422033] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.422033] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.422033] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] vm_ref = self.build_virtual_machine(instance, [ 585.422298] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.422298] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.422298] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.422298] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] for vif in network_info: [ 585.422298] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 585.422298] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] return self._sync_wrapper(fn, *args, **kwargs) [ 585.422298] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 585.422298] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] self.wait() [ 585.422298] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 585.422298] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] self[:] = self._gt.wait() [ 585.422298] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.422298] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] return self._exit_event.wait() [ 585.422298] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.422555] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] result = hub.switch() [ 585.422555] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.422555] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] return self.greenlet.switch() [ 585.422555] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.422555] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] result = function(*args, **kwargs) [ 585.422555] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.422555] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] return func(*args, **kwargs) [ 585.422555] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.422555] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] raise e [ 585.422555] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.422555] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] nwinfo = self.network_api.allocate_for_instance( [ 585.422555] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.422555] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] created_port_ids = self._update_ports_for_instance( [ 585.422807] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.422807] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] with excutils.save_and_reraise_exception(): [ 585.422807] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.422807] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] self.force_reraise() [ 585.422807] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.422807] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] raise self.value [ 585.422807] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.422807] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] updated_port = self._update_port( [ 585.422807] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.422807] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] _ensure_no_port_binding_failure(port) [ 585.422807] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.422807] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] raise exception.PortBindingFailed(port_id=port['id']) [ 585.423099] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] nova.exception.PortBindingFailed: Binding failed for port 5a8a05c3-4601-48bc-a242-f4627a6e0d1a, please check neutron logs for more information. [ 585.423099] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] [ 585.423099] env[61867]: INFO nova.compute.manager [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Terminating instance [ 585.427497] env[61867]: DEBUG oslo_concurrency.lockutils [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquiring lock "refresh_cache-af4b928f-25bc-4eff-9ba9-58afeae887f2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.427497] env[61867]: DEBUG oslo_concurrency.lockutils [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquired lock "refresh_cache-af4b928f-25bc-4eff-9ba9-58afeae887f2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.427497] env[61867]: DEBUG nova.network.neutron [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 585.570535] env[61867]: DEBUG nova.scheduler.client.report [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 585.678652] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Acquiring lock "69d292d9-2330-47ac-94d4-6797abb8c167" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.679297] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Lock "69d292d9-2330-47ac-94d4-6797abb8c167" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.707345] env[61867]: DEBUG nova.network.neutron [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.958569] env[61867]: DEBUG nova.network.neutron [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.080023] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.082s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.081794] env[61867]: DEBUG oslo_concurrency.lockutils [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.137s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.091510] env[61867]: DEBUG nova.network.neutron [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.112546] env[61867]: INFO nova.scheduler.client.report [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Deleted allocations for instance 76d3a228-e314-42b1-9eec-43a32fbd8035 [ 586.140569] env[61867]: DEBUG nova.compute.manager [req-f03c2048-7e4a-4407-b3da-aa4313b6ed97 req-a850b532-98f6-4288-a21b-e2ed71f65b9d service nova] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Received event network-changed-5a8a05c3-4601-48bc-a242-f4627a6e0d1a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 586.140880] env[61867]: DEBUG nova.compute.manager [req-f03c2048-7e4a-4407-b3da-aa4313b6ed97 req-a850b532-98f6-4288-a21b-e2ed71f65b9d service nova] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Refreshing instance network info cache due to event network-changed-5a8a05c3-4601-48bc-a242-f4627a6e0d1a. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 586.140947] env[61867]: DEBUG oslo_concurrency.lockutils [req-f03c2048-7e4a-4407-b3da-aa4313b6ed97 req-a850b532-98f6-4288-a21b-e2ed71f65b9d service nova] Acquiring lock "refresh_cache-af4b928f-25bc-4eff-9ba9-58afeae887f2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.209550] env[61867]: INFO nova.compute.manager [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] [instance: 11292ffa-0db5-4d70-a3c1-31b81bf4182e] Took 1.03 seconds to deallocate network for instance. [ 586.596897] env[61867]: DEBUG oslo_concurrency.lockutils [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Releasing lock "refresh_cache-af4b928f-25bc-4eff-9ba9-58afeae887f2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.597579] env[61867]: DEBUG nova.compute.manager [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 586.597579] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 586.598127] env[61867]: DEBUG oslo_concurrency.lockutils [req-f03c2048-7e4a-4407-b3da-aa4313b6ed97 req-a850b532-98f6-4288-a21b-e2ed71f65b9d service nova] Acquired lock "refresh_cache-af4b928f-25bc-4eff-9ba9-58afeae887f2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.598390] env[61867]: DEBUG nova.network.neutron [req-f03c2048-7e4a-4407-b3da-aa4313b6ed97 req-a850b532-98f6-4288-a21b-e2ed71f65b9d service nova] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Refreshing network info cache for port 5a8a05c3-4601-48bc-a242-f4627a6e0d1a {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 586.599751] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-17736cf1-357a-4fe2-885d-463fb5c39dd6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.609431] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2403b9da-aa37-4b01-b2a9-5de6152899ec {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.626581] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8246db49-4259-4b6b-bab4-149799b9cae5 tempest-ServerDiagnosticsV248Test-914366985 tempest-ServerDiagnosticsV248Test-914366985-project-member] Lock "76d3a228-e314-42b1-9eec-43a32fbd8035" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.453s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.638855] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance af4b928f-25bc-4eff-9ba9-58afeae887f2 could not be found. [ 586.639191] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 586.639377] env[61867]: INFO nova.compute.manager [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 586.639636] env[61867]: DEBUG oslo.service.loopingcall [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 586.639901] env[61867]: DEBUG nova.compute.manager [-] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 586.640031] env[61867]: DEBUG nova.network.neutron [-] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 586.656152] env[61867]: DEBUG nova.network.neutron [-] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.058305] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006c9f06-f889-4094-91d3-d39d35549e58 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.066181] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a969c56-c177-4fca-91da-1f4b2e87a35a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.099170] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5331f02c-6f0c-4e0f-9464-e3491a6750be {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.107335] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05f40c3-3efd-4418-ab0c-431ca0e406e2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.121600] env[61867]: DEBUG nova.compute.provider_tree [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.123444] env[61867]: DEBUG nova.network.neutron [req-f03c2048-7e4a-4407-b3da-aa4313b6ed97 req-a850b532-98f6-4288-a21b-e2ed71f65b9d service nova] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.158018] env[61867]: DEBUG nova.network.neutron [-] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.245887] env[61867]: DEBUG nova.network.neutron [req-f03c2048-7e4a-4407-b3da-aa4313b6ed97 req-a850b532-98f6-4288-a21b-e2ed71f65b9d service nova] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.248059] env[61867]: INFO nova.scheduler.client.report [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Deleted allocations for instance 11292ffa-0db5-4d70-a3c1-31b81bf4182e [ 587.388037] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquiring lock "3f1696b0-4de4-4128-bc83-e539e48dc8e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.388910] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "3f1696b0-4de4-4128-bc83-e539e48dc8e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.416315] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquiring lock "2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.416731] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.445948] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquiring lock "add3bbbb-fd30-49ad-b95e-601af5d790e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.446445] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "add3bbbb-fd30-49ad-b95e-601af5d790e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.626851] env[61867]: DEBUG nova.scheduler.client.report [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 587.661517] env[61867]: INFO nova.compute.manager [-] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Took 1.02 seconds to deallocate network for instance. [ 587.664946] env[61867]: DEBUG nova.compute.claims [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 587.665189] env[61867]: DEBUG oslo_concurrency.lockutils [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.755219] env[61867]: DEBUG oslo_concurrency.lockutils [req-f03c2048-7e4a-4407-b3da-aa4313b6ed97 req-a850b532-98f6-4288-a21b-e2ed71f65b9d service nova] Releasing lock "refresh_cache-af4b928f-25bc-4eff-9ba9-58afeae887f2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.759021] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f64f8fa7-08d6-4bce-abd0-96d5a583f39d tempest-ServerMetadataNegativeTestJSON-949715407 tempest-ServerMetadataNegativeTestJSON-949715407-project-member] Lock "11292ffa-0db5-4d70-a3c1-31b81bf4182e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.648s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.133402] env[61867]: DEBUG oslo_concurrency.lockutils [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.053s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.136663] env[61867]: ERROR nova.compute.manager [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b9b17023-6b11-4ffe-9df6-23d94c67b75b, please check neutron logs for more information. [ 588.136663] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Traceback (most recent call last): [ 588.136663] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 588.136663] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] self.driver.spawn(context, instance, image_meta, [ 588.136663] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 588.136663] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.136663] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.136663] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] vm_ref = self.build_virtual_machine(instance, [ 588.136663] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.136663] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.136663] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.136931] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] for vif in network_info: [ 588.136931] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.136931] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] return self._sync_wrapper(fn, *args, **kwargs) [ 588.136931] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.136931] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] self.wait() [ 588.136931] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.136931] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] self[:] = self._gt.wait() [ 588.136931] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.136931] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] return self._exit_event.wait() [ 588.136931] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 588.136931] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] result = hub.switch() [ 588.136931] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 588.136931] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] return self.greenlet.switch() [ 588.137277] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.137277] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] result = function(*args, **kwargs) [ 588.137277] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.137277] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] return func(*args, **kwargs) [ 588.137277] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.137277] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] raise e [ 588.137277] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.137277] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] nwinfo = self.network_api.allocate_for_instance( [ 588.137277] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.137277] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] created_port_ids = self._update_ports_for_instance( [ 588.137277] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.137277] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] with excutils.save_and_reraise_exception(): [ 588.137277] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.137573] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] self.force_reraise() [ 588.137573] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.137573] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] raise self.value [ 588.137573] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.137573] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] updated_port = self._update_port( [ 588.137573] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.137573] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] _ensure_no_port_binding_failure(port) [ 588.137573] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.137573] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] raise exception.PortBindingFailed(port_id=port['id']) [ 588.137573] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] nova.exception.PortBindingFailed: Binding failed for port b9b17023-6b11-4ffe-9df6-23d94c67b75b, please check neutron logs for more information. [ 588.137573] env[61867]: ERROR nova.compute.manager [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] [ 588.137823] env[61867]: DEBUG nova.compute.utils [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Binding failed for port b9b17023-6b11-4ffe-9df6-23d94c67b75b, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 588.137823] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.112s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.139289] env[61867]: INFO nova.compute.claims [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 588.142712] env[61867]: DEBUG nova.compute.manager [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Build of instance f911465c-277a-41ee-a6b5-0e3b85185b56 was re-scheduled: Binding failed for port b9b17023-6b11-4ffe-9df6-23d94c67b75b, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 588.142869] env[61867]: DEBUG nova.compute.manager [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 588.143228] env[61867]: DEBUG oslo_concurrency.lockutils [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Acquiring lock "refresh_cache-f911465c-277a-41ee-a6b5-0e3b85185b56" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.143286] env[61867]: DEBUG oslo_concurrency.lockutils [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Acquired lock "refresh_cache-f911465c-277a-41ee-a6b5-0e3b85185b56" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.143409] env[61867]: DEBUG nova.network.neutron [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.265872] env[61867]: DEBUG nova.compute.manager [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 588.635718] env[61867]: DEBUG nova.compute.manager [req-7bb032d5-db34-4edc-bb6a-d9f3e986bf45 req-72b28701-8946-4867-9c81-f53a7accfc03 service nova] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Received event network-vif-deleted-5a8a05c3-4601-48bc-a242-f4627a6e0d1a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 588.666491] env[61867]: DEBUG nova.network.neutron [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.759620] env[61867]: DEBUG nova.network.neutron [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.428614] env[61867]: DEBUG oslo_concurrency.lockutils [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Releasing lock "refresh_cache-f911465c-277a-41ee-a6b5-0e3b85185b56" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.428614] env[61867]: DEBUG nova.compute.manager [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 589.428614] env[61867]: DEBUG nova.compute.manager [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.428880] env[61867]: DEBUG nova.network.neutron [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 589.450810] env[61867]: DEBUG oslo_concurrency.lockutils [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.458792] env[61867]: DEBUG nova.network.neutron [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.941841] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f90ccf8-a603-4212-9749-5cd28945969c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.951941] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed250385-5a76-4564-a54c-24179c6d044b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.985009] env[61867]: DEBUG nova.network.neutron [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.986898] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302c4218-e908-4f06-9a9f-9a07b70014e7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.996092] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-007d764d-7b10-4fb9-a314-2754cf0ee8d5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.010298] env[61867]: DEBUG nova.compute.provider_tree [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.102119] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Acquiring lock "5dab4fcf-1a5e-493a-842b-b652e99d67ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.102241] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Lock "5dab4fcf-1a5e-493a-842b-b652e99d67ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.490711] env[61867]: INFO nova.compute.manager [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] [instance: f911465c-277a-41ee-a6b5-0e3b85185b56] Took 1.06 seconds to deallocate network for instance. [ 590.512787] env[61867]: DEBUG nova.scheduler.client.report [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 591.020017] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.883s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.020705] env[61867]: DEBUG nova.compute.manager [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 591.024167] env[61867]: DEBUG oslo_concurrency.lockutils [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.394s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.530883] env[61867]: DEBUG nova.compute.utils [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 591.539403] env[61867]: INFO nova.scheduler.client.report [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Deleted allocations for instance f911465c-277a-41ee-a6b5-0e3b85185b56 [ 591.546834] env[61867]: DEBUG nova.compute.manager [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 591.547033] env[61867]: DEBUG nova.network.neutron [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 591.603334] env[61867]: DEBUG nova.policy [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '624eae2c3c6f49e09bee42d57e40be2e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3234e3e4e57549d6a6af972a05fb7448', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 591.984804] env[61867]: DEBUG nova.network.neutron [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Successfully created port: b376247a-ff0d-4244-a596-81b52382e3d1 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 592.047960] env[61867]: DEBUG nova.compute.manager [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 592.054243] env[61867]: DEBUG oslo_concurrency.lockutils [None req-53ef7fb3-9c0b-4f14-9b64-58beace7e34d tempest-ServerActionsTestJSON-1382160688 tempest-ServerActionsTestJSON-1382160688-project-member] Lock "f911465c-277a-41ee-a6b5-0e3b85185b56" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.167s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.087952] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde37463-f15e-4a83-9876-1ecf410e2a25 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.098123] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f05a0e36-de26-48a7-a0a2-0d46c4442259 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.131536] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb327634-d40f-4093-8427-8f0db4d9c434 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.140086] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36b0925b-05c0-4929-a7b7-fef574699144 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.156054] env[61867]: DEBUG nova.compute.provider_tree [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.519249] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "ab18091f-bab5-4926-b7a0-1f0e005bebe0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.519844] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "ab18091f-bab5-4926-b7a0-1f0e005bebe0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.560887] env[61867]: DEBUG nova.compute.manager [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 592.658578] env[61867]: DEBUG nova.scheduler.client.report [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 593.065553] env[61867]: DEBUG nova.compute.manager [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 593.105343] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.105343] env[61867]: DEBUG nova.virt.hardware [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 593.108032] env[61867]: DEBUG nova.virt.hardware [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 593.108032] env[61867]: DEBUG nova.virt.hardware [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 593.108032] env[61867]: DEBUG nova.virt.hardware [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 593.108032] env[61867]: DEBUG nova.virt.hardware [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 593.108032] env[61867]: DEBUG nova.virt.hardware [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 593.108223] env[61867]: DEBUG nova.virt.hardware [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 593.108223] env[61867]: DEBUG nova.virt.hardware [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 593.108393] env[61867]: DEBUG nova.virt.hardware [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 593.108580] env[61867]: DEBUG nova.virt.hardware [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 593.108771] env[61867]: DEBUG nova.virt.hardware [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 593.109884] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9a6079a-7a7a-45d1-b29e-b6337067d0a4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.118951] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184450d4-9f90-4f10-914a-a2c5bc179d74 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.164344] env[61867]: DEBUG oslo_concurrency.lockutils [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.140s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.165050] env[61867]: ERROR nova.compute.manager [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 425c84f1-a9a7-4602-b422-a01b089bab3e, please check neutron logs for more information. [ 593.165050] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Traceback (most recent call last): [ 593.165050] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 593.165050] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] self.driver.spawn(context, instance, image_meta, [ 593.165050] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 593.165050] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 593.165050] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 593.165050] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] vm_ref = self.build_virtual_machine(instance, [ 593.165050] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 593.165050] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] vif_infos = vmwarevif.get_vif_info(self._session, [ 593.165050] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 593.165355] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] for vif in network_info: [ 593.165355] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 593.165355] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] return self._sync_wrapper(fn, *args, **kwargs) [ 593.165355] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 593.165355] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] self.wait() [ 593.165355] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 593.165355] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] self[:] = self._gt.wait() [ 593.165355] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 593.165355] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] return self._exit_event.wait() [ 593.165355] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 593.165355] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] result = hub.switch() [ 593.165355] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 593.165355] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] return self.greenlet.switch() [ 593.165631] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.165631] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] result = function(*args, **kwargs) [ 593.165631] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 593.165631] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] return func(*args, **kwargs) [ 593.165631] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.165631] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] raise e [ 593.165631] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.165631] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] nwinfo = self.network_api.allocate_for_instance( [ 593.165631] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.165631] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] created_port_ids = self._update_ports_for_instance( [ 593.165631] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.165631] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] with excutils.save_and_reraise_exception(): [ 593.165631] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.165903] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] self.force_reraise() [ 593.165903] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.165903] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] raise self.value [ 593.165903] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.165903] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] updated_port = self._update_port( [ 593.165903] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.165903] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] _ensure_no_port_binding_failure(port) [ 593.165903] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.165903] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] raise exception.PortBindingFailed(port_id=port['id']) [ 593.165903] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] nova.exception.PortBindingFailed: Binding failed for port 425c84f1-a9a7-4602-b422-a01b089bab3e, please check neutron logs for more information. [ 593.165903] env[61867]: ERROR nova.compute.manager [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] [ 593.166151] env[61867]: DEBUG nova.compute.utils [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Binding failed for port 425c84f1-a9a7-4602-b422-a01b089bab3e, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 593.170375] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.447s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.170936] env[61867]: DEBUG nova.compute.manager [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Build of instance 814dd6d2-43b1-4700-b585-c5d33b96931f was re-scheduled: Binding failed for port 425c84f1-a9a7-4602-b422-a01b089bab3e, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 593.171482] env[61867]: DEBUG nova.compute.manager [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 593.171734] env[61867]: DEBUG oslo_concurrency.lockutils [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Acquiring lock "refresh_cache-814dd6d2-43b1-4700-b585-c5d33b96931f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.171905] env[61867]: DEBUG oslo_concurrency.lockutils [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Acquired lock "refresh_cache-814dd6d2-43b1-4700-b585-c5d33b96931f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.172101] env[61867]: DEBUG nova.network.neutron [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 593.684658] env[61867]: ERROR nova.compute.manager [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b376247a-ff0d-4244-a596-81b52382e3d1, please check neutron logs for more information. [ 593.684658] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 593.684658] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.684658] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 593.684658] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.684658] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 593.684658] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.684658] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 593.684658] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.684658] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 593.684658] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.684658] env[61867]: ERROR nova.compute.manager raise self.value [ 593.684658] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.684658] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 593.684658] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.684658] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 593.685041] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.685041] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 593.685041] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b376247a-ff0d-4244-a596-81b52382e3d1, please check neutron logs for more information. [ 593.685041] env[61867]: ERROR nova.compute.manager [ 593.685041] env[61867]: Traceback (most recent call last): [ 593.685041] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 593.685041] env[61867]: listener.cb(fileno) [ 593.685041] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.685041] env[61867]: result = function(*args, **kwargs) [ 593.685041] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 593.685041] env[61867]: return func(*args, **kwargs) [ 593.685296] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.685296] env[61867]: raise e [ 593.685296] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.685296] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 593.685296] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.685296] env[61867]: created_port_ids = self._update_ports_for_instance( [ 593.685296] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.685296] env[61867]: with excutils.save_and_reraise_exception(): [ 593.685296] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.685296] env[61867]: self.force_reraise() [ 593.685296] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.685296] env[61867]: raise self.value [ 593.685296] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.685296] env[61867]: updated_port = self._update_port( [ 593.685296] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.685296] env[61867]: _ensure_no_port_binding_failure(port) [ 593.685296] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.685296] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 593.685296] env[61867]: nova.exception.PortBindingFailed: Binding failed for port b376247a-ff0d-4244-a596-81b52382e3d1, please check neutron logs for more information. [ 593.685296] env[61867]: Removing descriptor: 21 [ 593.685793] env[61867]: ERROR nova.compute.manager [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b376247a-ff0d-4244-a596-81b52382e3d1, please check neutron logs for more information. [ 593.685793] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Traceback (most recent call last): [ 593.685793] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 593.685793] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] yield resources [ 593.685793] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 593.685793] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] self.driver.spawn(context, instance, image_meta, [ 593.685793] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 593.685793] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] self._vmops.spawn(context, instance, image_meta, injected_files, [ 593.685793] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 593.685793] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] vm_ref = self.build_virtual_machine(instance, [ 593.685793] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 593.686077] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] vif_infos = vmwarevif.get_vif_info(self._session, [ 593.686077] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 593.686077] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] for vif in network_info: [ 593.686077] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 593.686077] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] return self._sync_wrapper(fn, *args, **kwargs) [ 593.686077] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 593.686077] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] self.wait() [ 593.686077] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 593.686077] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] self[:] = self._gt.wait() [ 593.686077] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 593.686077] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] return self._exit_event.wait() [ 593.686077] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 593.686077] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] result = hub.switch() [ 593.686454] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 593.686454] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] return self.greenlet.switch() [ 593.686454] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.686454] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] result = function(*args, **kwargs) [ 593.686454] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 593.686454] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] return func(*args, **kwargs) [ 593.686454] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.686454] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] raise e [ 593.686454] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.686454] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] nwinfo = self.network_api.allocate_for_instance( [ 593.686454] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.686454] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] created_port_ids = self._update_ports_for_instance( [ 593.686454] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.686732] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] with excutils.save_and_reraise_exception(): [ 593.686732] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.686732] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] self.force_reraise() [ 593.686732] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.686732] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] raise self.value [ 593.686732] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.686732] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] updated_port = self._update_port( [ 593.686732] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.686732] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] _ensure_no_port_binding_failure(port) [ 593.686732] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.686732] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] raise exception.PortBindingFailed(port_id=port['id']) [ 593.686732] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] nova.exception.PortBindingFailed: Binding failed for port b376247a-ff0d-4244-a596-81b52382e3d1, please check neutron logs for more information. [ 593.686732] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] [ 593.687049] env[61867]: INFO nova.compute.manager [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Terminating instance [ 593.690895] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquiring lock "refresh_cache-45e2004d-731b-4002-84b2-3f9138c09663" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.690895] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquired lock "refresh_cache-45e2004d-731b-4002-84b2-3f9138c09663" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.690895] env[61867]: DEBUG nova.network.neutron [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 593.697444] env[61867]: DEBUG nova.network.neutron [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.753884] env[61867]: DEBUG nova.network.neutron [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.837885] env[61867]: DEBUG nova.compute.manager [req-831050cd-37e5-4beb-8828-7cd254494aa2 req-d65320df-6a67-4ec4-89ca-c37900771dfb service nova] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Received event network-changed-b376247a-ff0d-4244-a596-81b52382e3d1 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 593.837885] env[61867]: DEBUG nova.compute.manager [req-831050cd-37e5-4beb-8828-7cd254494aa2 req-d65320df-6a67-4ec4-89ca-c37900771dfb service nova] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Refreshing instance network info cache due to event network-changed-b376247a-ff0d-4244-a596-81b52382e3d1. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 593.837885] env[61867]: DEBUG oslo_concurrency.lockutils [req-831050cd-37e5-4beb-8828-7cd254494aa2 req-d65320df-6a67-4ec4-89ca-c37900771dfb service nova] Acquiring lock "refresh_cache-45e2004d-731b-4002-84b2-3f9138c09663" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.202879] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1145e564-d2e6-4d2b-af5f-d5040c73637e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.209460] env[61867]: DEBUG nova.network.neutron [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.211802] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751a329b-2f64-4e23-9332-9211eb6443fc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.242642] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-def867c7-e18a-4509-a79e-d547bb35e80c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.250402] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e57ba49-a54c-449f-8faa-aec3bea83ac1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.266425] env[61867]: DEBUG oslo_concurrency.lockutils [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Releasing lock "refresh_cache-814dd6d2-43b1-4700-b585-c5d33b96931f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.266582] env[61867]: DEBUG nova.compute.manager [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 594.267028] env[61867]: DEBUG nova.compute.manager [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 594.267028] env[61867]: DEBUG nova.network.neutron [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 594.268894] env[61867]: DEBUG nova.compute.provider_tree [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.306883] env[61867]: DEBUG nova.network.neutron [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.328618] env[61867]: DEBUG nova.network.neutron [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.775234] env[61867]: DEBUG nova.scheduler.client.report [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.811229] env[61867]: DEBUG nova.network.neutron [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.832768] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Releasing lock "refresh_cache-45e2004d-731b-4002-84b2-3f9138c09663" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.832768] env[61867]: DEBUG nova.compute.manager [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 594.832768] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 594.833322] env[61867]: DEBUG oslo_concurrency.lockutils [req-831050cd-37e5-4beb-8828-7cd254494aa2 req-d65320df-6a67-4ec4-89ca-c37900771dfb service nova] Acquired lock "refresh_cache-45e2004d-731b-4002-84b2-3f9138c09663" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.833384] env[61867]: DEBUG nova.network.neutron [req-831050cd-37e5-4beb-8828-7cd254494aa2 req-d65320df-6a67-4ec4-89ca-c37900771dfb service nova] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Refreshing network info cache for port b376247a-ff0d-4244-a596-81b52382e3d1 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 594.834507] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-981db8b8-4222-4834-b265-12d6f37ee326 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.848923] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7634f4-3699-4f77-a747-faafae378f33 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.878351] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 45e2004d-731b-4002-84b2-3f9138c09663 could not be found. [ 594.878351] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 594.878351] env[61867]: INFO nova.compute.manager [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Took 0.04 seconds to destroy the instance on the hypervisor. [ 594.878351] env[61867]: DEBUG oslo.service.loopingcall [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 594.878351] env[61867]: DEBUG nova.compute.manager [-] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 594.878351] env[61867]: DEBUG nova.network.neutron [-] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 594.899965] env[61867]: DEBUG nova.network.neutron [-] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.281120] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.113s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.282487] env[61867]: ERROR nova.compute.manager [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port baaff768-f304-4adb-a80f-87f99daa6bfd, please check neutron logs for more information. [ 595.282487] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Traceback (most recent call last): [ 595.282487] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 595.282487] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] self.driver.spawn(context, instance, image_meta, [ 595.282487] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 595.282487] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] self._vmops.spawn(context, instance, image_meta, injected_files, [ 595.282487] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 595.282487] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] vm_ref = self.build_virtual_machine(instance, [ 595.282487] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 595.282487] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] vif_infos = vmwarevif.get_vif_info(self._session, [ 595.282487] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 595.283358] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] for vif in network_info: [ 595.283358] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 595.283358] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] return self._sync_wrapper(fn, *args, **kwargs) [ 595.283358] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 595.283358] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] self.wait() [ 595.283358] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 595.283358] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] self[:] = self._gt.wait() [ 595.283358] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 595.283358] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] return self._exit_event.wait() [ 595.283358] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 595.283358] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] result = hub.switch() [ 595.283358] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 595.283358] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] return self.greenlet.switch() [ 595.283646] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.283646] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] result = function(*args, **kwargs) [ 595.283646] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.283646] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] return func(*args, **kwargs) [ 595.283646] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.283646] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] raise e [ 595.283646] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.283646] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] nwinfo = self.network_api.allocate_for_instance( [ 595.283646] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.283646] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] created_port_ids = self._update_ports_for_instance( [ 595.283646] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.283646] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] with excutils.save_and_reraise_exception(): [ 595.283646] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.283933] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] self.force_reraise() [ 595.283933] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.283933] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] raise self.value [ 595.283933] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.283933] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] updated_port = self._update_port( [ 595.283933] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.283933] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] _ensure_no_port_binding_failure(port) [ 595.283933] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.283933] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] raise exception.PortBindingFailed(port_id=port['id']) [ 595.283933] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] nova.exception.PortBindingFailed: Binding failed for port baaff768-f304-4adb-a80f-87f99daa6bfd, please check neutron logs for more information. [ 595.283933] env[61867]: ERROR nova.compute.manager [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] [ 595.284239] env[61867]: DEBUG nova.compute.utils [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Binding failed for port baaff768-f304-4adb-a80f-87f99daa6bfd, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 595.284239] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.692s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.285491] env[61867]: INFO nova.compute.claims [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 595.288746] env[61867]: DEBUG nova.compute.manager [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Build of instance 0d716b01-a5d5-40f6-b002-5e1488bd54af was re-scheduled: Binding failed for port baaff768-f304-4adb-a80f-87f99daa6bfd, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 595.289215] env[61867]: DEBUG nova.compute.manager [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 595.289434] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Acquiring lock "refresh_cache-0d716b01-a5d5-40f6-b002-5e1488bd54af" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.289578] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Acquired lock "refresh_cache-0d716b01-a5d5-40f6-b002-5e1488bd54af" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.289905] env[61867]: DEBUG nova.network.neutron [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 595.313698] env[61867]: INFO nova.compute.manager [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] [instance: 814dd6d2-43b1-4700-b585-c5d33b96931f] Took 1.05 seconds to deallocate network for instance. [ 595.360820] env[61867]: DEBUG nova.network.neutron [req-831050cd-37e5-4beb-8828-7cd254494aa2 req-d65320df-6a67-4ec4-89ca-c37900771dfb service nova] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.392506] env[61867]: DEBUG oslo_concurrency.lockutils [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "b1ec32c5-0642-4dca-ad43-05a5172d04d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.392896] env[61867]: DEBUG oslo_concurrency.lockutils [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "b1ec32c5-0642-4dca-ad43-05a5172d04d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.402271] env[61867]: DEBUG nova.network.neutron [-] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.542944] env[61867]: DEBUG nova.network.neutron [req-831050cd-37e5-4beb-8828-7cd254494aa2 req-d65320df-6a67-4ec4-89ca-c37900771dfb service nova] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.824150] env[61867]: DEBUG nova.network.neutron [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.908730] env[61867]: INFO nova.compute.manager [-] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Took 1.03 seconds to deallocate network for instance. [ 595.911664] env[61867]: DEBUG nova.compute.claims [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 595.911787] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.988730] env[61867]: DEBUG nova.network.neutron [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.046926] env[61867]: DEBUG oslo_concurrency.lockutils [req-831050cd-37e5-4beb-8828-7cd254494aa2 req-d65320df-6a67-4ec4-89ca-c37900771dfb service nova] Releasing lock "refresh_cache-45e2004d-731b-4002-84b2-3f9138c09663" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.052920] env[61867]: DEBUG nova.compute.manager [req-035df8c6-1b1b-4b3f-b1e9-d485ef890f51 req-ed2891dc-f096-4ce9-a97b-589c8f0c9d9e service nova] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Received event network-vif-deleted-b376247a-ff0d-4244-a596-81b52382e3d1 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 596.099526] env[61867]: DEBUG oslo_concurrency.lockutils [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "8656c3b8-b7b1-442a-9719-1588ee83d19b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.100097] env[61867]: DEBUG oslo_concurrency.lockutils [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "8656c3b8-b7b1-442a-9719-1588ee83d19b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.356398] env[61867]: INFO nova.scheduler.client.report [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Deleted allocations for instance 814dd6d2-43b1-4700-b585-c5d33b96931f [ 596.492136] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Releasing lock "refresh_cache-0d716b01-a5d5-40f6-b002-5e1488bd54af" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.492441] env[61867]: DEBUG nova.compute.manager [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 596.492441] env[61867]: DEBUG nova.compute.manager [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 596.492621] env[61867]: DEBUG nova.network.neutron [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 596.514338] env[61867]: DEBUG nova.network.neutron [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.799120] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0343dd-1fec-4174-a609-0f46cc789a78 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.806920] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f747a9-07f6-403c-84a7-e88849a84f39 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.837608] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e522e4-4b57-484a-806f-2089ac771670 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.845524] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28465dac-a387-4ac4-adfa-2f1be292433d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.859962] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 596.860447] env[61867]: DEBUG nova.compute.provider_tree [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.861589] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 596.868183] env[61867]: DEBUG oslo_concurrency.lockutils [None req-857383f9-5e2e-4f8e-8566-2b7f1c8ddb23 tempest-DeleteServersAdminTestJSON-1087276998 tempest-DeleteServersAdminTestJSON-1087276998-project-member] Lock "814dd6d2-43b1-4700-b585-c5d33b96931f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.459s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.016835] env[61867]: DEBUG nova.network.neutron [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.366396] env[61867]: DEBUG nova.scheduler.client.report [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.370925] env[61867]: DEBUG nova.compute.manager [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 597.378408] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.378408] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Starting heal instance info cache {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 597.378408] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Rebuilding the list of instances to heal {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 597.521658] env[61867]: INFO nova.compute.manager [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] [instance: 0d716b01-a5d5-40f6-b002-5e1488bd54af] Took 1.03 seconds to deallocate network for instance. [ 597.875414] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.591s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.877028] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 597.880242] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.322s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.882363] env[61867]: INFO nova.compute.claims [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 597.890215] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 597.890413] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 597.890549] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 597.890682] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 597.890808] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Didn't find any instances for network info cache update. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 597.891226] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.891397] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.891547] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.891696] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.891837] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.891980] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.896023] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61867) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 597.897262] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.922225] env[61867]: DEBUG oslo_concurrency.lockutils [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.393497] env[61867]: DEBUG nova.compute.utils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 598.394781] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 598.394955] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 598.400160] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.556808] env[61867]: INFO nova.scheduler.client.report [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Deleted allocations for instance 0d716b01-a5d5-40f6-b002-5e1488bd54af [ 598.579269] env[61867]: DEBUG nova.policy [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9781b7a61f15483b8577f854266a876d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8451dfe23754279965eda6b2f2f37cf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 598.900358] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 599.071973] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db2dae88-515a-4284-ba04-8107e22d9615 tempest-ServerTagsTestJSON-1139575384 tempest-ServerTagsTestJSON-1139575384-project-member] Lock "0d716b01-a5d5-40f6-b002-5e1488bd54af" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.629s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.078259] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Successfully created port: 3ca41fad-6498-4d3f-ac59-22f2e90bcbbe {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 599.362068] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad63b0b9-7680-49b9-be85-08bcdd938e16 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.372512] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca473aec-d75c-4453-b031-28b40c3b1fb5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.406659] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af945828-d514-461f-aff3-25179d42e165 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.418137] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fdde5b2-6137-4412-a6e7-d3cf466ba4d2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.434947] env[61867]: DEBUG nova.compute.provider_tree [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.574583] env[61867]: DEBUG nova.compute.manager [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 599.916077] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 599.936930] env[61867]: DEBUG nova.scheduler.client.report [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.943047] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 599.943306] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 599.943472] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 599.943740] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 599.943800] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 599.943952] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 599.944179] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 599.944348] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 599.944520] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 599.944689] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 599.944870] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 599.945769] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae244506-9a00-412d-9cf2-3304325af52e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.960728] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b9f67e-0f2c-4624-a8dd-80fca993e4fc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.105114] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.357455] env[61867]: DEBUG nova.compute.manager [req-573f7bac-b75c-480b-8b48-a0a79e58692d req-3f9cc132-f8a5-4e3c-aacf-90946e537f5f service nova] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Received event network-changed-3ca41fad-6498-4d3f-ac59-22f2e90bcbbe {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 600.357686] env[61867]: DEBUG nova.compute.manager [req-573f7bac-b75c-480b-8b48-a0a79e58692d req-3f9cc132-f8a5-4e3c-aacf-90946e537f5f service nova] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Refreshing instance network info cache due to event network-changed-3ca41fad-6498-4d3f-ac59-22f2e90bcbbe. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 600.358241] env[61867]: DEBUG oslo_concurrency.lockutils [req-573f7bac-b75c-480b-8b48-a0a79e58692d req-3f9cc132-f8a5-4e3c-aacf-90946e537f5f service nova] Acquiring lock "refresh_cache-2ae534f8-1921-4892-8556-e2732624f412" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.360303] env[61867]: DEBUG oslo_concurrency.lockutils [req-573f7bac-b75c-480b-8b48-a0a79e58692d req-3f9cc132-f8a5-4e3c-aacf-90946e537f5f service nova] Acquired lock "refresh_cache-2ae534f8-1921-4892-8556-e2732624f412" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.360578] env[61867]: DEBUG nova.network.neutron [req-573f7bac-b75c-480b-8b48-a0a79e58692d req-3f9cc132-f8a5-4e3c-aacf-90946e537f5f service nova] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Refreshing network info cache for port 3ca41fad-6498-4d3f-ac59-22f2e90bcbbe {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 600.457982] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.575s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.457982] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 600.460395] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.768s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.524053] env[61867]: ERROR nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3ca41fad-6498-4d3f-ac59-22f2e90bcbbe, please check neutron logs for more information. [ 600.524053] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 600.524053] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.524053] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 600.524053] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.524053] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 600.524053] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.524053] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 600.524053] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.524053] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 600.524053] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.524053] env[61867]: ERROR nova.compute.manager raise self.value [ 600.524053] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.524053] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 600.524053] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.524053] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 600.524503] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.524503] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 600.524503] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3ca41fad-6498-4d3f-ac59-22f2e90bcbbe, please check neutron logs for more information. [ 600.524503] env[61867]: ERROR nova.compute.manager [ 600.524503] env[61867]: Traceback (most recent call last): [ 600.524503] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 600.524503] env[61867]: listener.cb(fileno) [ 600.524503] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.524503] env[61867]: result = function(*args, **kwargs) [ 600.524503] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 600.524503] env[61867]: return func(*args, **kwargs) [ 600.524503] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.524503] env[61867]: raise e [ 600.524503] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.524503] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 600.524503] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.524503] env[61867]: created_port_ids = self._update_ports_for_instance( [ 600.524503] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.524503] env[61867]: with excutils.save_and_reraise_exception(): [ 600.524503] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.524503] env[61867]: self.force_reraise() [ 600.524503] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.524503] env[61867]: raise self.value [ 600.524503] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.524503] env[61867]: updated_port = self._update_port( [ 600.524503] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.524503] env[61867]: _ensure_no_port_binding_failure(port) [ 600.524503] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.524503] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 600.525258] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 3ca41fad-6498-4d3f-ac59-22f2e90bcbbe, please check neutron logs for more information. [ 600.525258] env[61867]: Removing descriptor: 17 [ 600.525258] env[61867]: ERROR nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3ca41fad-6498-4d3f-ac59-22f2e90bcbbe, please check neutron logs for more information. [ 600.525258] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] Traceback (most recent call last): [ 600.525258] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 600.525258] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] yield resources [ 600.525258] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 600.525258] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] self.driver.spawn(context, instance, image_meta, [ 600.525258] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 600.525258] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.525258] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.525258] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] vm_ref = self.build_virtual_machine(instance, [ 600.525519] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.525519] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.525519] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.525519] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] for vif in network_info: [ 600.525519] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 600.525519] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] return self._sync_wrapper(fn, *args, **kwargs) [ 600.525519] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 600.525519] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] self.wait() [ 600.525519] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 600.525519] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] self[:] = self._gt.wait() [ 600.525519] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.525519] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] return self._exit_event.wait() [ 600.525519] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 600.525793] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] result = hub.switch() [ 600.525793] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 600.525793] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] return self.greenlet.switch() [ 600.525793] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.525793] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] result = function(*args, **kwargs) [ 600.525793] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 600.525793] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] return func(*args, **kwargs) [ 600.525793] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.525793] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] raise e [ 600.525793] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.525793] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] nwinfo = self.network_api.allocate_for_instance( [ 600.525793] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.525793] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] created_port_ids = self._update_ports_for_instance( [ 600.526075] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.526075] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] with excutils.save_and_reraise_exception(): [ 600.526075] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.526075] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] self.force_reraise() [ 600.526075] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.526075] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] raise self.value [ 600.526075] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.526075] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] updated_port = self._update_port( [ 600.526075] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.526075] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] _ensure_no_port_binding_failure(port) [ 600.526075] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.526075] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] raise exception.PortBindingFailed(port_id=port['id']) [ 600.526402] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] nova.exception.PortBindingFailed: Binding failed for port 3ca41fad-6498-4d3f-ac59-22f2e90bcbbe, please check neutron logs for more information. [ 600.526402] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] [ 600.526402] env[61867]: INFO nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Terminating instance [ 600.531083] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "refresh_cache-2ae534f8-1921-4892-8556-e2732624f412" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.799437] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Acquiring lock "5e97f142-b5c0-4318-9a35-085234d1cb64" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.799838] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Lock "5e97f142-b5c0-4318-9a35-085234d1cb64" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.967426] env[61867]: DEBUG nova.compute.utils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 600.976137] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 600.976333] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 601.049115] env[61867]: DEBUG nova.network.neutron [req-573f7bac-b75c-480b-8b48-a0a79e58692d req-3f9cc132-f8a5-4e3c-aacf-90946e537f5f service nova] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.149269] env[61867]: DEBUG nova.policy [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9781b7a61f15483b8577f854266a876d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8451dfe23754279965eda6b2f2f37cf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 601.296639] env[61867]: DEBUG nova.network.neutron [req-573f7bac-b75c-480b-8b48-a0a79e58692d req-3f9cc132-f8a5-4e3c-aacf-90946e537f5f service nova] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.470967] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c011774-fc43-4abe-bc65-9504704d221f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.476456] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 601.482570] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d7bcca3-298f-4324-8abc-c0c841413811 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.524134] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01715aaf-e90c-47b1-ac42-42f833d5885e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.533182] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c11681-c7ab-4179-ba72-9ced8423d9cd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.549200] env[61867]: DEBUG nova.compute.provider_tree [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.801915] env[61867]: DEBUG oslo_concurrency.lockutils [req-573f7bac-b75c-480b-8b48-a0a79e58692d req-3f9cc132-f8a5-4e3c-aacf-90946e537f5f service nova] Releasing lock "refresh_cache-2ae534f8-1921-4892-8556-e2732624f412" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.801915] env[61867]: DEBUG oslo_concurrency.lockutils [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Acquiring lock "796ab4d3-ea9a-495d-9807-7780b1a7b0d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.801915] env[61867]: DEBUG oslo_concurrency.lockutils [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Lock "796ab4d3-ea9a-495d-9807-7780b1a7b0d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.801915] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquired lock "refresh_cache-2ae534f8-1921-4892-8556-e2732624f412" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.802162] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 601.845780] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Successfully created port: 8c5f6b14-4a4d-49f9-a67a-99aadc4b4211 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 602.051940] env[61867]: DEBUG nova.scheduler.client.report [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.344915] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.437370] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Acquiring lock "8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.437370] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Lock "8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.442521] env[61867]: DEBUG nova.compute.manager [req-35441882-7853-4a82-a774-0486a362d65e req-167496bd-c6da-49aa-a387-0bbb6d1ada2f service nova] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Received event network-vif-deleted-3ca41fad-6498-4d3f-ac59-22f2e90bcbbe {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 602.488505] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 602.529859] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 602.529859] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 602.529859] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.530144] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 602.530144] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.530144] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 602.530144] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 602.530144] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 602.530339] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 602.530339] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 602.530339] env[61867]: DEBUG nova.virt.hardware [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 602.530339] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1980a5d2-b8a2-48e8-b064-1fb9a7c66559 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.542819] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d092e410-2e93-453c-818b-c9adb25c0b69 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.559186] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.101s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.561697] env[61867]: ERROR nova.compute.manager [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 007ca2b7-ee1b-4705-94e9-e47938ebe5bf, please check neutron logs for more information. [ 602.561697] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Traceback (most recent call last): [ 602.561697] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.561697] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] self.driver.spawn(context, instance, image_meta, [ 602.561697] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 602.561697] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.561697] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.561697] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] vm_ref = self.build_virtual_machine(instance, [ 602.561697] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.561697] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.561697] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.562170] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] for vif in network_info: [ 602.562170] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.562170] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] return self._sync_wrapper(fn, *args, **kwargs) [ 602.562170] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.562170] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] self.wait() [ 602.562170] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.562170] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] self[:] = self._gt.wait() [ 602.562170] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.562170] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] return self._exit_event.wait() [ 602.562170] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.562170] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] result = hub.switch() [ 602.562170] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.562170] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] return self.greenlet.switch() [ 602.562622] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.562622] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] result = function(*args, **kwargs) [ 602.562622] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.562622] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] return func(*args, **kwargs) [ 602.562622] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.562622] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] raise e [ 602.562622] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.562622] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] nwinfo = self.network_api.allocate_for_instance( [ 602.562622] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.562622] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] created_port_ids = self._update_ports_for_instance( [ 602.562622] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.562622] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] with excutils.save_and_reraise_exception(): [ 602.562622] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.562972] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] self.force_reraise() [ 602.562972] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.562972] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] raise self.value [ 602.562972] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.562972] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] updated_port = self._update_port( [ 602.562972] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.562972] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] _ensure_no_port_binding_failure(port) [ 602.562972] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.562972] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] raise exception.PortBindingFailed(port_id=port['id']) [ 602.562972] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] nova.exception.PortBindingFailed: Binding failed for port 007ca2b7-ee1b-4705-94e9-e47938ebe5bf, please check neutron logs for more information. [ 602.562972] env[61867]: ERROR nova.compute.manager [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] [ 602.563275] env[61867]: DEBUG nova.compute.utils [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Binding failed for port 007ca2b7-ee1b-4705-94e9-e47938ebe5bf, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 602.563912] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.433s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.565465] env[61867]: INFO nova.compute.claims [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 602.568970] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.569882] env[61867]: DEBUG nova.compute.manager [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Build of instance 76c297c0-f393-499a-80e5-a78d84ec0885 was re-scheduled: Binding failed for port 007ca2b7-ee1b-4705-94e9-e47938ebe5bf, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 602.570412] env[61867]: DEBUG nova.compute.manager [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 602.570642] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquiring lock "refresh_cache-76c297c0-f393-499a-80e5-a78d84ec0885" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.570785] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquired lock "refresh_cache-76c297c0-f393-499a-80e5-a78d84ec0885" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.571073] env[61867]: DEBUG nova.network.neutron [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 603.073167] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Releasing lock "refresh_cache-2ae534f8-1921-4892-8556-e2732624f412" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.073597] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 603.073783] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 603.078270] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3eac7f65-5c24-4d89-9f22-bdc8c4466a8b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.090421] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ef33a8-13c0-47a9-a6ea-b00a1b1afd01 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.116892] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2ae534f8-1921-4892-8556-e2732624f412 could not be found. [ 603.117145] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 603.118300] env[61867]: INFO nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Took 0.04 seconds to destroy the instance on the hypervisor. [ 603.118300] env[61867]: DEBUG oslo.service.loopingcall [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 603.118300] env[61867]: DEBUG nova.compute.manager [-] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.118300] env[61867]: DEBUG nova.network.neutron [-] [instance: 2ae534f8-1921-4892-8556-e2732624f412] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 603.131021] env[61867]: DEBUG nova.network.neutron [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.150677] env[61867]: DEBUG nova.network.neutron [-] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.356840] env[61867]: DEBUG nova.network.neutron [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.653449] env[61867]: DEBUG nova.network.neutron [-] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.864042] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Releasing lock "refresh_cache-76c297c0-f393-499a-80e5-a78d84ec0885" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.864290] env[61867]: DEBUG nova.compute.manager [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 603.864469] env[61867]: DEBUG nova.compute.manager [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.864658] env[61867]: DEBUG nova.network.neutron [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 603.893061] env[61867]: DEBUG nova.network.neutron [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.065816] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9bfbea6-ab5a-4d24-a4ef-61b4326408fc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.073456] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f174c0a-880e-4a7d-b47e-4b72eb88b202 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.108904] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e11588-4cf9-4ccf-8904-f61a314ef10c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.116028] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6598894-d002-47ec-a896-0aeb2483a040 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.129723] env[61867]: DEBUG nova.compute.provider_tree [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.144222] env[61867]: ERROR nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8c5f6b14-4a4d-49f9-a67a-99aadc4b4211, please check neutron logs for more information. [ 604.144222] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 604.144222] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.144222] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 604.144222] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.144222] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 604.144222] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.144222] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 604.144222] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.144222] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 604.144222] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.144222] env[61867]: ERROR nova.compute.manager raise self.value [ 604.144222] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.144222] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 604.144222] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.144222] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 604.144588] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.144588] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 604.144588] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8c5f6b14-4a4d-49f9-a67a-99aadc4b4211, please check neutron logs for more information. [ 604.144588] env[61867]: ERROR nova.compute.manager [ 604.144588] env[61867]: Traceback (most recent call last): [ 604.144588] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 604.144588] env[61867]: listener.cb(fileno) [ 604.144588] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.144588] env[61867]: result = function(*args, **kwargs) [ 604.144588] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.144588] env[61867]: return func(*args, **kwargs) [ 604.144588] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.144588] env[61867]: raise e [ 604.144588] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.144588] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 604.144588] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.144588] env[61867]: created_port_ids = self._update_ports_for_instance( [ 604.144588] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.144588] env[61867]: with excutils.save_and_reraise_exception(): [ 604.144588] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.144588] env[61867]: self.force_reraise() [ 604.144588] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.144588] env[61867]: raise self.value [ 604.144588] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.144588] env[61867]: updated_port = self._update_port( [ 604.144588] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.144588] env[61867]: _ensure_no_port_binding_failure(port) [ 604.144588] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.144588] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 604.145295] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 8c5f6b14-4a4d-49f9-a67a-99aadc4b4211, please check neutron logs for more information. [ 604.145295] env[61867]: Removing descriptor: 21 [ 604.145492] env[61867]: ERROR nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8c5f6b14-4a4d-49f9-a67a-99aadc4b4211, please check neutron logs for more information. [ 604.145492] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Traceback (most recent call last): [ 604.145492] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 604.145492] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] yield resources [ 604.145492] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.145492] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] self.driver.spawn(context, instance, image_meta, [ 604.145492] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 604.145492] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.145492] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.145492] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] vm_ref = self.build_virtual_machine(instance, [ 604.145492] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.145772] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.145772] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.145772] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] for vif in network_info: [ 604.145772] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.145772] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] return self._sync_wrapper(fn, *args, **kwargs) [ 604.145772] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.145772] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] self.wait() [ 604.145772] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.145772] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] self[:] = self._gt.wait() [ 604.145772] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.145772] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] return self._exit_event.wait() [ 604.145772] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.145772] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] result = hub.switch() [ 604.146074] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.146074] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] return self.greenlet.switch() [ 604.146074] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.146074] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] result = function(*args, **kwargs) [ 604.146074] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.146074] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] return func(*args, **kwargs) [ 604.146074] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.146074] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] raise e [ 604.146074] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.146074] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] nwinfo = self.network_api.allocate_for_instance( [ 604.146074] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.146074] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] created_port_ids = self._update_ports_for_instance( [ 604.146074] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.146359] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] with excutils.save_and_reraise_exception(): [ 604.146359] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.146359] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] self.force_reraise() [ 604.146359] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.146359] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] raise self.value [ 604.146359] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.146359] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] updated_port = self._update_port( [ 604.146359] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.146359] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] _ensure_no_port_binding_failure(port) [ 604.146359] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.146359] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] raise exception.PortBindingFailed(port_id=port['id']) [ 604.146359] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] nova.exception.PortBindingFailed: Binding failed for port 8c5f6b14-4a4d-49f9-a67a-99aadc4b4211, please check neutron logs for more information. [ 604.146359] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] [ 604.146643] env[61867]: INFO nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Terminating instance [ 604.148883] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "refresh_cache-18b98634-d76d-496b-904f-1a9c289cdb8a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.148883] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquired lock "refresh_cache-18b98634-d76d-496b-904f-1a9c289cdb8a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.148883] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 604.156199] env[61867]: INFO nova.compute.manager [-] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Took 1.04 seconds to deallocate network for instance. [ 604.160549] env[61867]: DEBUG nova.compute.claims [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 604.160549] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.399083] env[61867]: DEBUG nova.network.neutron [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.445468] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "311a8d4f-68dc-4e0c-acdb-d3929afb9a1b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.446037] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "311a8d4f-68dc-4e0c-acdb-d3929afb9a1b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.475665] env[61867]: DEBUG nova.compute.manager [req-6ff978fb-450d-4729-bac0-6506a737cf38 req-4538ab97-17f6-4a99-addd-abc6b880e4e3 service nova] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Received event network-changed-8c5f6b14-4a4d-49f9-a67a-99aadc4b4211 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 604.475860] env[61867]: DEBUG nova.compute.manager [req-6ff978fb-450d-4729-bac0-6506a737cf38 req-4538ab97-17f6-4a99-addd-abc6b880e4e3 service nova] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Refreshing instance network info cache due to event network-changed-8c5f6b14-4a4d-49f9-a67a-99aadc4b4211. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 604.476055] env[61867]: DEBUG oslo_concurrency.lockutils [req-6ff978fb-450d-4729-bac0-6506a737cf38 req-4538ab97-17f6-4a99-addd-abc6b880e4e3 service nova] Acquiring lock "refresh_cache-18b98634-d76d-496b-904f-1a9c289cdb8a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.633423] env[61867]: DEBUG nova.scheduler.client.report [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.667122] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.746688] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.902760] env[61867]: INFO nova.compute.manager [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 76c297c0-f393-499a-80e5-a78d84ec0885] Took 1.04 seconds to deallocate network for instance. [ 605.138737] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.574s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.139337] env[61867]: DEBUG nova.compute.manager [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 605.143663] env[61867]: DEBUG oslo_concurrency.lockutils [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.478s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.253136] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Releasing lock "refresh_cache-18b98634-d76d-496b-904f-1a9c289cdb8a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.253726] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 605.253990] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 605.254626] env[61867]: DEBUG oslo_concurrency.lockutils [req-6ff978fb-450d-4729-bac0-6506a737cf38 req-4538ab97-17f6-4a99-addd-abc6b880e4e3 service nova] Acquired lock "refresh_cache-18b98634-d76d-496b-904f-1a9c289cdb8a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.254886] env[61867]: DEBUG nova.network.neutron [req-6ff978fb-450d-4729-bac0-6506a737cf38 req-4538ab97-17f6-4a99-addd-abc6b880e4e3 service nova] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Refreshing network info cache for port 8c5f6b14-4a4d-49f9-a67a-99aadc4b4211 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 605.256040] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-17e12131-4eb9-4119-af21-ccba13557564 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.266954] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b03b1de5-e893-43dd-9d6f-060552f6d5a9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.292688] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 18b98634-d76d-496b-904f-1a9c289cdb8a could not be found. [ 605.292920] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 605.293112] env[61867]: INFO nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 605.293353] env[61867]: DEBUG oslo.service.loopingcall [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 605.293595] env[61867]: DEBUG nova.compute.manager [-] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.293687] env[61867]: DEBUG nova.network.neutron [-] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 605.314516] env[61867]: DEBUG nova.network.neutron [-] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.648564] env[61867]: DEBUG nova.compute.utils [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 605.657152] env[61867]: DEBUG nova.compute.manager [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 605.657152] env[61867]: DEBUG nova.network.neutron [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 605.699663] env[61867]: DEBUG nova.policy [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cbe42936a2b44fc4b60bfd6d228b6559', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '191bfd0c17cf43259932ef658596430e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 605.778581] env[61867]: DEBUG nova.network.neutron [req-6ff978fb-450d-4729-bac0-6506a737cf38 req-4538ab97-17f6-4a99-addd-abc6b880e4e3 service nova] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.816209] env[61867]: DEBUG nova.network.neutron [-] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.900218] env[61867]: DEBUG nova.network.neutron [req-6ff978fb-450d-4729-bac0-6506a737cf38 req-4538ab97-17f6-4a99-addd-abc6b880e4e3 service nova] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.947788] env[61867]: INFO nova.scheduler.client.report [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Deleted allocations for instance 76c297c0-f393-499a-80e5-a78d84ec0885 [ 606.111590] env[61867]: DEBUG nova.network.neutron [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Successfully created port: 06047bd3-a41f-45f1-9e3f-7b129fd2a39a {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 606.154986] env[61867]: DEBUG nova.compute.manager [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 606.199220] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08601379-833d-4201-9947-e4ccf6e457bf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.210357] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39112557-4acc-40c4-8a29-39cdfca06fb7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.256358] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d4e07c-2937-4c8d-9172-8a0b47959d49 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.265468] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5feb67d-805e-4da0-9ec8-634749925a33 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.281922] env[61867]: DEBUG nova.compute.provider_tree [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.319058] env[61867]: INFO nova.compute.manager [-] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Took 1.03 seconds to deallocate network for instance. [ 606.321594] env[61867]: DEBUG nova.compute.claims [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 606.321784] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.404133] env[61867]: DEBUG oslo_concurrency.lockutils [req-6ff978fb-450d-4729-bac0-6506a737cf38 req-4538ab97-17f6-4a99-addd-abc6b880e4e3 service nova] Releasing lock "refresh_cache-18b98634-d76d-496b-904f-1a9c289cdb8a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.404405] env[61867]: DEBUG nova.compute.manager [req-6ff978fb-450d-4729-bac0-6506a737cf38 req-4538ab97-17f6-4a99-addd-abc6b880e4e3 service nova] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Received event network-vif-deleted-8c5f6b14-4a4d-49f9-a67a-99aadc4b4211 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 606.465173] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2e916b85-532f-4341-83f0-c72f4356c75d tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "76c297c0-f393-499a-80e5-a78d84ec0885" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.191s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.787355] env[61867]: DEBUG nova.scheduler.client.report [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.968514] env[61867]: DEBUG nova.compute.manager [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 607.108668] env[61867]: ERROR nova.compute.manager [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 06047bd3-a41f-45f1-9e3f-7b129fd2a39a, please check neutron logs for more information. [ 607.108668] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 607.108668] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.108668] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 607.108668] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.108668] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 607.108668] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.108668] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 607.108668] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.108668] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 607.108668] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.108668] env[61867]: ERROR nova.compute.manager raise self.value [ 607.108668] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.108668] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 607.108668] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.108668] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 607.109100] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.109100] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 607.109100] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 06047bd3-a41f-45f1-9e3f-7b129fd2a39a, please check neutron logs for more information. [ 607.109100] env[61867]: ERROR nova.compute.manager [ 607.109100] env[61867]: Traceback (most recent call last): [ 607.109100] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 607.109100] env[61867]: listener.cb(fileno) [ 607.109100] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.109100] env[61867]: result = function(*args, **kwargs) [ 607.109100] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.109100] env[61867]: return func(*args, **kwargs) [ 607.109100] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.109100] env[61867]: raise e [ 607.109100] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.109100] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 607.109100] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.109100] env[61867]: created_port_ids = self._update_ports_for_instance( [ 607.109100] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.109100] env[61867]: with excutils.save_and_reraise_exception(): [ 607.109100] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.109100] env[61867]: self.force_reraise() [ 607.109100] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.109100] env[61867]: raise self.value [ 607.109100] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.109100] env[61867]: updated_port = self._update_port( [ 607.109100] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.109100] env[61867]: _ensure_no_port_binding_failure(port) [ 607.109100] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.109100] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 607.109708] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 06047bd3-a41f-45f1-9e3f-7b129fd2a39a, please check neutron logs for more information. [ 607.109708] env[61867]: Removing descriptor: 21 [ 607.125128] env[61867]: DEBUG nova.compute.manager [req-6f7cda41-44e5-481b-b3fa-6d7443d5b44f req-8e610bb6-0d6e-4aa7-9327-9527687ce10d service nova] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Received event network-changed-06047bd3-a41f-45f1-9e3f-7b129fd2a39a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 607.125382] env[61867]: DEBUG nova.compute.manager [req-6f7cda41-44e5-481b-b3fa-6d7443d5b44f req-8e610bb6-0d6e-4aa7-9327-9527687ce10d service nova] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Refreshing instance network info cache due to event network-changed-06047bd3-a41f-45f1-9e3f-7b129fd2a39a. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 607.125574] env[61867]: DEBUG oslo_concurrency.lockutils [req-6f7cda41-44e5-481b-b3fa-6d7443d5b44f req-8e610bb6-0d6e-4aa7-9327-9527687ce10d service nova] Acquiring lock "refresh_cache-736f861f-d9a5-4fa5-a28d-26604ebbf051" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.125700] env[61867]: DEBUG oslo_concurrency.lockutils [req-6f7cda41-44e5-481b-b3fa-6d7443d5b44f req-8e610bb6-0d6e-4aa7-9327-9527687ce10d service nova] Acquired lock "refresh_cache-736f861f-d9a5-4fa5-a28d-26604ebbf051" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.125859] env[61867]: DEBUG nova.network.neutron [req-6f7cda41-44e5-481b-b3fa-6d7443d5b44f req-8e610bb6-0d6e-4aa7-9327-9527687ce10d service nova] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Refreshing network info cache for port 06047bd3-a41f-45f1-9e3f-7b129fd2a39a {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 607.169996] env[61867]: DEBUG nova.compute.manager [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 607.198220] env[61867]: DEBUG nova.virt.hardware [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 607.198517] env[61867]: DEBUG nova.virt.hardware [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 607.198718] env[61867]: DEBUG nova.virt.hardware [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 607.198913] env[61867]: DEBUG nova.virt.hardware [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 607.199121] env[61867]: DEBUG nova.virt.hardware [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 607.199287] env[61867]: DEBUG nova.virt.hardware [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 607.199524] env[61867]: DEBUG nova.virt.hardware [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 607.199686] env[61867]: DEBUG nova.virt.hardware [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 607.199858] env[61867]: DEBUG nova.virt.hardware [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 607.200604] env[61867]: DEBUG nova.virt.hardware [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 607.200604] env[61867]: DEBUG nova.virt.hardware [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 607.201769] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76c29dc8-b747-4d39-9fd3-053ad831b427 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.211315] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3f40e0-c91f-4e9f-9555-f38d11e3d5ea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.233017] env[61867]: ERROR nova.compute.manager [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 06047bd3-a41f-45f1-9e3f-7b129fd2a39a, please check neutron logs for more information. [ 607.233017] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Traceback (most recent call last): [ 607.233017] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 607.233017] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] yield resources [ 607.233017] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 607.233017] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] self.driver.spawn(context, instance, image_meta, [ 607.233017] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 607.233017] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.233017] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.233017] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] vm_ref = self.build_virtual_machine(instance, [ 607.233017] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.233361] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.233361] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.233361] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] for vif in network_info: [ 607.233361] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.233361] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] return self._sync_wrapper(fn, *args, **kwargs) [ 607.233361] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.233361] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] self.wait() [ 607.233361] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.233361] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] self[:] = self._gt.wait() [ 607.233361] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.233361] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] return self._exit_event.wait() [ 607.233361] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 607.233361] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] current.throw(*self._exc) [ 607.233855] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.233855] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] result = function(*args, **kwargs) [ 607.233855] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.233855] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] return func(*args, **kwargs) [ 607.233855] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.233855] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] raise e [ 607.233855] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.233855] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] nwinfo = self.network_api.allocate_for_instance( [ 607.233855] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.233855] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] created_port_ids = self._update_ports_for_instance( [ 607.233855] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.233855] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] with excutils.save_and_reraise_exception(): [ 607.233855] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.234257] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] self.force_reraise() [ 607.234257] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.234257] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] raise self.value [ 607.234257] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.234257] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] updated_port = self._update_port( [ 607.234257] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.234257] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] _ensure_no_port_binding_failure(port) [ 607.234257] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.234257] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] raise exception.PortBindingFailed(port_id=port['id']) [ 607.234257] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] nova.exception.PortBindingFailed: Binding failed for port 06047bd3-a41f-45f1-9e3f-7b129fd2a39a, please check neutron logs for more information. [ 607.234257] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] [ 607.234257] env[61867]: INFO nova.compute.manager [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Terminating instance [ 607.236512] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Acquiring lock "refresh_cache-736f861f-d9a5-4fa5-a28d-26604ebbf051" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.293581] env[61867]: DEBUG oslo_concurrency.lockutils [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.150s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.294231] env[61867]: ERROR nova.compute.manager [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5a8a05c3-4601-48bc-a242-f4627a6e0d1a, please check neutron logs for more information. [ 607.294231] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Traceback (most recent call last): [ 607.294231] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 607.294231] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] self.driver.spawn(context, instance, image_meta, [ 607.294231] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 607.294231] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.294231] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.294231] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] vm_ref = self.build_virtual_machine(instance, [ 607.294231] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.294231] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.294231] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.294678] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] for vif in network_info: [ 607.294678] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.294678] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] return self._sync_wrapper(fn, *args, **kwargs) [ 607.294678] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.294678] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] self.wait() [ 607.294678] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.294678] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] self[:] = self._gt.wait() [ 607.294678] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.294678] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] return self._exit_event.wait() [ 607.294678] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 607.294678] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] result = hub.switch() [ 607.294678] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 607.294678] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] return self.greenlet.switch() [ 607.295061] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.295061] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] result = function(*args, **kwargs) [ 607.295061] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.295061] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] return func(*args, **kwargs) [ 607.295061] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.295061] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] raise e [ 607.295061] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.295061] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] nwinfo = self.network_api.allocate_for_instance( [ 607.295061] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.295061] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] created_port_ids = self._update_ports_for_instance( [ 607.295061] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.295061] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] with excutils.save_and_reraise_exception(): [ 607.295061] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.295620] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] self.force_reraise() [ 607.295620] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.295620] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] raise self.value [ 607.295620] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.295620] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] updated_port = self._update_port( [ 607.295620] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.295620] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] _ensure_no_port_binding_failure(port) [ 607.295620] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.295620] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] raise exception.PortBindingFailed(port_id=port['id']) [ 607.295620] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] nova.exception.PortBindingFailed: Binding failed for port 5a8a05c3-4601-48bc-a242-f4627a6e0d1a, please check neutron logs for more information. [ 607.295620] env[61867]: ERROR nova.compute.manager [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] [ 607.295871] env[61867]: DEBUG nova.compute.utils [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Binding failed for port 5a8a05c3-4601-48bc-a242-f4627a6e0d1a, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 607.296314] env[61867]: DEBUG oslo_concurrency.lockutils [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.846s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.298153] env[61867]: INFO nova.compute.claims [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 607.302045] env[61867]: DEBUG nova.compute.manager [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Build of instance af4b928f-25bc-4eff-9ba9-58afeae887f2 was re-scheduled: Binding failed for port 5a8a05c3-4601-48bc-a242-f4627a6e0d1a, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 607.302045] env[61867]: DEBUG nova.compute.manager [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 607.302045] env[61867]: DEBUG oslo_concurrency.lockutils [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquiring lock "refresh_cache-af4b928f-25bc-4eff-9ba9-58afeae887f2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.302045] env[61867]: DEBUG oslo_concurrency.lockutils [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquired lock "refresh_cache-af4b928f-25bc-4eff-9ba9-58afeae887f2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.302242] env[61867]: DEBUG nova.network.neutron [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 607.491205] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.645015] env[61867]: DEBUG nova.network.neutron [req-6f7cda41-44e5-481b-b3fa-6d7443d5b44f req-8e610bb6-0d6e-4aa7-9327-9527687ce10d service nova] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.718287] env[61867]: DEBUG nova.network.neutron [req-6f7cda41-44e5-481b-b3fa-6d7443d5b44f req-8e610bb6-0d6e-4aa7-9327-9527687ce10d service nova] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.821577] env[61867]: DEBUG nova.network.neutron [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.909561] env[61867]: DEBUG nova.network.neutron [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.221239] env[61867]: DEBUG oslo_concurrency.lockutils [req-6f7cda41-44e5-481b-b3fa-6d7443d5b44f req-8e610bb6-0d6e-4aa7-9327-9527687ce10d service nova] Releasing lock "refresh_cache-736f861f-d9a5-4fa5-a28d-26604ebbf051" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.221677] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Acquired lock "refresh_cache-736f861f-d9a5-4fa5-a28d-26604ebbf051" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.221867] env[61867]: DEBUG nova.network.neutron [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 608.411395] env[61867]: DEBUG oslo_concurrency.lockutils [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Releasing lock "refresh_cache-af4b928f-25bc-4eff-9ba9-58afeae887f2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.411698] env[61867]: DEBUG nova.compute.manager [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 608.411951] env[61867]: DEBUG nova.compute.manager [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.412169] env[61867]: DEBUG nova.network.neutron [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 608.454868] env[61867]: DEBUG nova.network.neutron [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.724860] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f160aca-d2d8-4e8e-a904-76cedc65004a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.734280] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3226b84-e60b-4bb6-a529-6bebb6c3c4c9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.763804] env[61867]: DEBUG nova.network.neutron [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.766010] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e4b44b-cf30-40b5-bca2-4a952bdd205f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.773214] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b634ae0-0be9-4664-88fc-98767f1f082b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.786525] env[61867]: DEBUG nova.compute.provider_tree [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.839604] env[61867]: DEBUG nova.network.neutron [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.962153] env[61867]: DEBUG nova.network.neutron [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.292019] env[61867]: DEBUG nova.scheduler.client.report [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.310124] env[61867]: DEBUG nova.compute.manager [req-c29c7222-1004-4ec9-a72d-8023b8f9046c req-b7030336-dad4-40c8-82c0-d156c81480bd service nova] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Received event network-vif-deleted-06047bd3-a41f-45f1-9e3f-7b129fd2a39a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 609.344958] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Releasing lock "refresh_cache-736f861f-d9a5-4fa5-a28d-26604ebbf051" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.344958] env[61867]: DEBUG nova.compute.manager [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 609.344958] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 609.345441] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ba033dab-1529-4707-801e-f56b5d48add5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.356046] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-156c59f1-7e24-4b24-943b-5d3841c2a8b1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.385254] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 736f861f-d9a5-4fa5-a28d-26604ebbf051 could not be found. [ 609.385254] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 609.385254] env[61867]: INFO nova.compute.manager [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Took 0.04 seconds to destroy the instance on the hypervisor. [ 609.385254] env[61867]: DEBUG oslo.service.loopingcall [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 609.385254] env[61867]: DEBUG nova.compute.manager [-] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 609.385254] env[61867]: DEBUG nova.network.neutron [-] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 609.404200] env[61867]: DEBUG nova.network.neutron [-] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.463626] env[61867]: INFO nova.compute.manager [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: af4b928f-25bc-4eff-9ba9-58afeae887f2] Took 1.05 seconds to deallocate network for instance. [ 609.795755] env[61867]: DEBUG oslo_concurrency.lockutils [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.796358] env[61867]: DEBUG nova.compute.manager [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 609.799214] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.697s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.800952] env[61867]: INFO nova.compute.claims [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 609.908311] env[61867]: DEBUG nova.network.neutron [-] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.313317] env[61867]: DEBUG nova.compute.utils [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 610.315880] env[61867]: DEBUG nova.compute.manager [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 610.316281] env[61867]: DEBUG nova.network.neutron [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 610.367739] env[61867]: DEBUG nova.policy [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8f6ff20a11ed4f608800be1c7931cf78', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e945237d2af146869ce3a23b223ad9fb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 610.410743] env[61867]: INFO nova.compute.manager [-] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Took 1.03 seconds to deallocate network for instance. [ 610.413141] env[61867]: DEBUG nova.compute.claims [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 610.413311] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.499740] env[61867]: INFO nova.scheduler.client.report [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Deleted allocations for instance af4b928f-25bc-4eff-9ba9-58afeae887f2 [ 610.758561] env[61867]: DEBUG nova.network.neutron [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Successfully created port: 984de365-ba36-4070-85a5-f2bbc4b223f6 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 610.821449] env[61867]: DEBUG nova.compute.manager [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 611.011787] env[61867]: DEBUG oslo_concurrency.lockutils [None req-73439605-89db-44c6-a740-e6203a941cdc tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "af4b928f-25bc-4eff-9ba9-58afeae887f2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.341s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.249149] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d29365d6-d0fd-44f1-8a92-ee5a65301d6a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.259250] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead06f6c-823a-41a7-bf6b-709036c55750 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.295555] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e603a9-7581-4e6f-b09f-35ac1bdefe68 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.303479] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a904b4-c585-4fa5-87a0-114c786657ba {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.320283] env[61867]: DEBUG nova.compute.provider_tree [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.515525] env[61867]: DEBUG nova.compute.manager [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 611.717325] env[61867]: DEBUG nova.compute.manager [req-8c857159-770d-422e-b3b0-30935519abda req-b715cdc7-2ffa-4dab-a6d0-66fcf3417392 service nova] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Received event network-changed-984de365-ba36-4070-85a5-f2bbc4b223f6 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 611.717325] env[61867]: DEBUG nova.compute.manager [req-8c857159-770d-422e-b3b0-30935519abda req-b715cdc7-2ffa-4dab-a6d0-66fcf3417392 service nova] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Refreshing instance network info cache due to event network-changed-984de365-ba36-4070-85a5-f2bbc4b223f6. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 611.717325] env[61867]: DEBUG oslo_concurrency.lockutils [req-8c857159-770d-422e-b3b0-30935519abda req-b715cdc7-2ffa-4dab-a6d0-66fcf3417392 service nova] Acquiring lock "refresh_cache-d25df58d-4b66-4aa4-9d07-205f36e61775" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.717325] env[61867]: DEBUG oslo_concurrency.lockutils [req-8c857159-770d-422e-b3b0-30935519abda req-b715cdc7-2ffa-4dab-a6d0-66fcf3417392 service nova] Acquired lock "refresh_cache-d25df58d-4b66-4aa4-9d07-205f36e61775" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.717325] env[61867]: DEBUG nova.network.neutron [req-8c857159-770d-422e-b3b0-30935519abda req-b715cdc7-2ffa-4dab-a6d0-66fcf3417392 service nova] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Refreshing network info cache for port 984de365-ba36-4070-85a5-f2bbc4b223f6 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 611.816515] env[61867]: ERROR nova.compute.manager [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 984de365-ba36-4070-85a5-f2bbc4b223f6, please check neutron logs for more information. [ 611.816515] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 611.816515] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.816515] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 611.816515] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.816515] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 611.816515] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.816515] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 611.816515] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.816515] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 611.816515] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.816515] env[61867]: ERROR nova.compute.manager raise self.value [ 611.816515] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.816515] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 611.816515] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.816515] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 611.817065] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.817065] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 611.817065] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 984de365-ba36-4070-85a5-f2bbc4b223f6, please check neutron logs for more information. [ 611.817065] env[61867]: ERROR nova.compute.manager [ 611.817065] env[61867]: Traceback (most recent call last): [ 611.817065] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 611.817065] env[61867]: listener.cb(fileno) [ 611.817065] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.817065] env[61867]: result = function(*args, **kwargs) [ 611.817065] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.817065] env[61867]: return func(*args, **kwargs) [ 611.817065] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.817065] env[61867]: raise e [ 611.817065] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.817065] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 611.817065] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.817065] env[61867]: created_port_ids = self._update_ports_for_instance( [ 611.817065] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.817065] env[61867]: with excutils.save_and_reraise_exception(): [ 611.817065] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.817065] env[61867]: self.force_reraise() [ 611.817065] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.817065] env[61867]: raise self.value [ 611.817065] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.817065] env[61867]: updated_port = self._update_port( [ 611.817065] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.817065] env[61867]: _ensure_no_port_binding_failure(port) [ 611.817065] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.817065] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 611.817893] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 984de365-ba36-4070-85a5-f2bbc4b223f6, please check neutron logs for more information. [ 611.817893] env[61867]: Removing descriptor: 21 [ 611.826900] env[61867]: DEBUG nova.scheduler.client.report [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.837905] env[61867]: DEBUG nova.compute.manager [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 611.865938] env[61867]: DEBUG nova.virt.hardware [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 611.866195] env[61867]: DEBUG nova.virt.hardware [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 611.866342] env[61867]: DEBUG nova.virt.hardware [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 611.866513] env[61867]: DEBUG nova.virt.hardware [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 611.866660] env[61867]: DEBUG nova.virt.hardware [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 611.866834] env[61867]: DEBUG nova.virt.hardware [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 611.867064] env[61867]: DEBUG nova.virt.hardware [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 611.867229] env[61867]: DEBUG nova.virt.hardware [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 611.867394] env[61867]: DEBUG nova.virt.hardware [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 611.867605] env[61867]: DEBUG nova.virt.hardware [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 611.867799] env[61867]: DEBUG nova.virt.hardware [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 611.868655] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8edaa02b-02f5-4b98-bbd3-caf482e95d3f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.879315] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad8559d-e43c-4f42-930c-e4566e50de8d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.893181] env[61867]: ERROR nova.compute.manager [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 984de365-ba36-4070-85a5-f2bbc4b223f6, please check neutron logs for more information. [ 611.893181] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Traceback (most recent call last): [ 611.893181] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 611.893181] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] yield resources [ 611.893181] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.893181] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] self.driver.spawn(context, instance, image_meta, [ 611.893181] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 611.893181] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.893181] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.893181] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] vm_ref = self.build_virtual_machine(instance, [ 611.893181] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.893508] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.893508] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.893508] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] for vif in network_info: [ 611.893508] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.893508] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] return self._sync_wrapper(fn, *args, **kwargs) [ 611.893508] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.893508] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] self.wait() [ 611.893508] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.893508] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] self[:] = self._gt.wait() [ 611.893508] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.893508] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] return self._exit_event.wait() [ 611.893508] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 611.893508] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] current.throw(*self._exc) [ 611.894285] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.894285] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] result = function(*args, **kwargs) [ 611.894285] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.894285] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] return func(*args, **kwargs) [ 611.894285] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.894285] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] raise e [ 611.894285] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.894285] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] nwinfo = self.network_api.allocate_for_instance( [ 611.894285] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.894285] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] created_port_ids = self._update_ports_for_instance( [ 611.894285] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.894285] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] with excutils.save_and_reraise_exception(): [ 611.894285] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.894907] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] self.force_reraise() [ 611.894907] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.894907] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] raise self.value [ 611.894907] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.894907] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] updated_port = self._update_port( [ 611.894907] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.894907] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] _ensure_no_port_binding_failure(port) [ 611.894907] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.894907] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] raise exception.PortBindingFailed(port_id=port['id']) [ 611.894907] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] nova.exception.PortBindingFailed: Binding failed for port 984de365-ba36-4070-85a5-f2bbc4b223f6, please check neutron logs for more information. [ 611.894907] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] [ 611.894907] env[61867]: INFO nova.compute.manager [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Terminating instance [ 611.896403] env[61867]: DEBUG oslo_concurrency.lockutils [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Acquiring lock "refresh_cache-d25df58d-4b66-4aa4-9d07-205f36e61775" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.038425] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.237918] env[61867]: DEBUG nova.network.neutron [req-8c857159-770d-422e-b3b0-30935519abda req-b715cdc7-2ffa-4dab-a6d0-66fcf3417392 service nova] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.331751] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.532s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.332607] env[61867]: DEBUG nova.compute.manager [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 612.335915] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.424s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.364740] env[61867]: DEBUG nova.network.neutron [req-8c857159-770d-422e-b3b0-30935519abda req-b715cdc7-2ffa-4dab-a6d0-66fcf3417392 service nova] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.846139] env[61867]: DEBUG nova.compute.utils [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 612.847673] env[61867]: DEBUG nova.compute.manager [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 612.848011] env[61867]: DEBUG nova.network.neutron [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 612.870348] env[61867]: DEBUG oslo_concurrency.lockutils [req-8c857159-770d-422e-b3b0-30935519abda req-b715cdc7-2ffa-4dab-a6d0-66fcf3417392 service nova] Releasing lock "refresh_cache-d25df58d-4b66-4aa4-9d07-205f36e61775" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.871192] env[61867]: DEBUG oslo_concurrency.lockutils [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Acquired lock "refresh_cache-d25df58d-4b66-4aa4-9d07-205f36e61775" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.871317] env[61867]: DEBUG nova.network.neutron [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.920761] env[61867]: DEBUG nova.policy [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5190495c45d640d19e2fa10df52582ca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6fb37213d58b438789aab75774bc9c0b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 613.336841] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e03533e7-ba40-4382-bdd9-df813bb9f0a0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.346036] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142d5fe1-a0cd-4051-b820-b942444e0cbf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.354312] env[61867]: DEBUG nova.compute.manager [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 613.393661] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fee61e1-ea79-449e-a453-2ecef937a8bd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.408409] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b75ffd10-d3f4-4133-8040-6cbcd5a93ef9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.425181] env[61867]: DEBUG nova.compute.provider_tree [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.434615] env[61867]: DEBUG nova.network.neutron [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.445621] env[61867]: DEBUG nova.network.neutron [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Successfully created port: 6d020051-3a6c-4f84-922b-85cfee5b7411 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 613.621860] env[61867]: DEBUG nova.network.neutron [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.782269] env[61867]: DEBUG nova.compute.manager [req-ac65e475-d843-4d50-91ec-db4601b17aad req-2470a65f-620a-4062-bea8-a1fa3dfbfbaf service nova] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Received event network-vif-deleted-984de365-ba36-4070-85a5-f2bbc4b223f6 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 613.930495] env[61867]: DEBUG nova.scheduler.client.report [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.126464] env[61867]: DEBUG oslo_concurrency.lockutils [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Releasing lock "refresh_cache-d25df58d-4b66-4aa4-9d07-205f36e61775" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.126708] env[61867]: DEBUG nova.compute.manager [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 614.127051] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 614.127219] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9878797d-dcc8-48fa-ae9a-fd79feecb06e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.136389] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e48b8cd-36a7-4a24-97a8-e30da37abd80 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.158618] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d25df58d-4b66-4aa4-9d07-205f36e61775 could not be found. [ 614.158851] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 614.159044] env[61867]: INFO nova.compute.manager [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Took 0.03 seconds to destroy the instance on the hypervisor. [ 614.159284] env[61867]: DEBUG oslo.service.loopingcall [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 614.159506] env[61867]: DEBUG nova.compute.manager [-] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.159599] env[61867]: DEBUG nova.network.neutron [-] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 614.185520] env[61867]: DEBUG nova.network.neutron [-] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.368338] env[61867]: DEBUG nova.compute.manager [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 614.396435] env[61867]: DEBUG nova.virt.hardware [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 614.396688] env[61867]: DEBUG nova.virt.hardware [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 614.396845] env[61867]: DEBUG nova.virt.hardware [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 614.397031] env[61867]: DEBUG nova.virt.hardware [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 614.397182] env[61867]: DEBUG nova.virt.hardware [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 614.397327] env[61867]: DEBUG nova.virt.hardware [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 614.397540] env[61867]: DEBUG nova.virt.hardware [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 614.398104] env[61867]: DEBUG nova.virt.hardware [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 614.398104] env[61867]: DEBUG nova.virt.hardware [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 614.398221] env[61867]: DEBUG nova.virt.hardware [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 614.398533] env[61867]: DEBUG nova.virt.hardware [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 614.399806] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc593cd9-dcd6-4112-828e-5f9a399a3902 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.407166] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7879eb2-78ed-4fb2-93a8-40e8028d9c07 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.434575] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.099s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.435220] env[61867]: ERROR nova.compute.manager [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b376247a-ff0d-4244-a596-81b52382e3d1, please check neutron logs for more information. [ 614.435220] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Traceback (most recent call last): [ 614.435220] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.435220] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] self.driver.spawn(context, instance, image_meta, [ 614.435220] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 614.435220] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.435220] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.435220] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] vm_ref = self.build_virtual_machine(instance, [ 614.435220] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.435220] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.435220] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.435628] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] for vif in network_info: [ 614.435628] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 614.435628] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] return self._sync_wrapper(fn, *args, **kwargs) [ 614.435628] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 614.435628] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] self.wait() [ 614.435628] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 614.435628] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] self[:] = self._gt.wait() [ 614.435628] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.435628] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] return self._exit_event.wait() [ 614.435628] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 614.435628] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] result = hub.switch() [ 614.435628] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 614.435628] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] return self.greenlet.switch() [ 614.435976] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.435976] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] result = function(*args, **kwargs) [ 614.435976] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.435976] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] return func(*args, **kwargs) [ 614.435976] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.435976] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] raise e [ 614.435976] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.435976] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] nwinfo = self.network_api.allocate_for_instance( [ 614.435976] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.435976] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] created_port_ids = self._update_ports_for_instance( [ 614.435976] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.435976] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] with excutils.save_and_reraise_exception(): [ 614.435976] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.436351] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] self.force_reraise() [ 614.436351] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.436351] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] raise self.value [ 614.436351] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.436351] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] updated_port = self._update_port( [ 614.436351] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.436351] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] _ensure_no_port_binding_failure(port) [ 614.436351] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.436351] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] raise exception.PortBindingFailed(port_id=port['id']) [ 614.436351] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] nova.exception.PortBindingFailed: Binding failed for port b376247a-ff0d-4244-a596-81b52382e3d1, please check neutron logs for more information. [ 614.436351] env[61867]: ERROR nova.compute.manager [instance: 45e2004d-731b-4002-84b2-3f9138c09663] [ 614.436660] env[61867]: DEBUG nova.compute.utils [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Binding failed for port b376247a-ff0d-4244-a596-81b52382e3d1, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 614.437157] env[61867]: DEBUG oslo_concurrency.lockutils [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.515s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.438776] env[61867]: INFO nova.compute.claims [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.441536] env[61867]: DEBUG nova.compute.manager [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Build of instance 45e2004d-731b-4002-84b2-3f9138c09663 was re-scheduled: Binding failed for port b376247a-ff0d-4244-a596-81b52382e3d1, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 614.441966] env[61867]: DEBUG nova.compute.manager [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 614.442202] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquiring lock "refresh_cache-45e2004d-731b-4002-84b2-3f9138c09663" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.442351] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Acquired lock "refresh_cache-45e2004d-731b-4002-84b2-3f9138c09663" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.442507] env[61867]: DEBUG nova.network.neutron [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 614.689793] env[61867]: DEBUG nova.network.neutron [-] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.696570] env[61867]: ERROR nova.compute.manager [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6d020051-3a6c-4f84-922b-85cfee5b7411, please check neutron logs for more information. [ 614.696570] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 614.696570] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.696570] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 614.696570] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.696570] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 614.696570] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.696570] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 614.696570] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.696570] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 614.696570] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.696570] env[61867]: ERROR nova.compute.manager raise self.value [ 614.696570] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.696570] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 614.696570] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.696570] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 614.697187] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.697187] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 614.697187] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6d020051-3a6c-4f84-922b-85cfee5b7411, please check neutron logs for more information. [ 614.697187] env[61867]: ERROR nova.compute.manager [ 614.697187] env[61867]: Traceback (most recent call last): [ 614.697187] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 614.697187] env[61867]: listener.cb(fileno) [ 614.697187] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.697187] env[61867]: result = function(*args, **kwargs) [ 614.697187] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.697187] env[61867]: return func(*args, **kwargs) [ 614.697187] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.697187] env[61867]: raise e [ 614.697187] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.697187] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 614.697187] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.697187] env[61867]: created_port_ids = self._update_ports_for_instance( [ 614.697187] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.697187] env[61867]: with excutils.save_and_reraise_exception(): [ 614.697187] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.697187] env[61867]: self.force_reraise() [ 614.697187] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.697187] env[61867]: raise self.value [ 614.697187] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.697187] env[61867]: updated_port = self._update_port( [ 614.697187] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.697187] env[61867]: _ensure_no_port_binding_failure(port) [ 614.697187] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.697187] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 614.698048] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 6d020051-3a6c-4f84-922b-85cfee5b7411, please check neutron logs for more information. [ 614.698048] env[61867]: Removing descriptor: 21 [ 614.698048] env[61867]: ERROR nova.compute.manager [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6d020051-3a6c-4f84-922b-85cfee5b7411, please check neutron logs for more information. [ 614.698048] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Traceback (most recent call last): [ 614.698048] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 614.698048] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] yield resources [ 614.698048] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.698048] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] self.driver.spawn(context, instance, image_meta, [ 614.698048] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 614.698048] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.698048] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.698048] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] vm_ref = self.build_virtual_machine(instance, [ 614.698436] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.698436] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.698436] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.698436] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] for vif in network_info: [ 614.698436] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 614.698436] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] return self._sync_wrapper(fn, *args, **kwargs) [ 614.698436] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 614.698436] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] self.wait() [ 614.698436] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 614.698436] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] self[:] = self._gt.wait() [ 614.698436] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.698436] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] return self._exit_event.wait() [ 614.698436] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 614.698820] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] result = hub.switch() [ 614.698820] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 614.698820] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] return self.greenlet.switch() [ 614.698820] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.698820] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] result = function(*args, **kwargs) [ 614.698820] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.698820] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] return func(*args, **kwargs) [ 614.698820] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.698820] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] raise e [ 614.698820] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.698820] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] nwinfo = self.network_api.allocate_for_instance( [ 614.698820] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.698820] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] created_port_ids = self._update_ports_for_instance( [ 614.699213] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.699213] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] with excutils.save_and_reraise_exception(): [ 614.699213] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.699213] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] self.force_reraise() [ 614.699213] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.699213] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] raise self.value [ 614.699213] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.699213] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] updated_port = self._update_port( [ 614.699213] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.699213] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] _ensure_no_port_binding_failure(port) [ 614.699213] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.699213] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] raise exception.PortBindingFailed(port_id=port['id']) [ 614.699515] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] nova.exception.PortBindingFailed: Binding failed for port 6d020051-3a6c-4f84-922b-85cfee5b7411, please check neutron logs for more information. [ 614.699515] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] [ 614.699515] env[61867]: INFO nova.compute.manager [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Terminating instance [ 614.701515] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Acquiring lock "refresh_cache-ba4cc401-7ae3-40bc-996c-13741a65d991" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.701515] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Acquired lock "refresh_cache-ba4cc401-7ae3-40bc-996c-13741a65d991" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.701515] env[61867]: DEBUG nova.network.neutron [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 614.964701] env[61867]: DEBUG nova.network.neutron [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.048677] env[61867]: DEBUG nova.network.neutron [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.194248] env[61867]: INFO nova.compute.manager [-] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Took 1.03 seconds to deallocate network for instance. [ 615.197697] env[61867]: DEBUG nova.compute.claims [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 615.197697] env[61867]: DEBUG oslo_concurrency.lockutils [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.220748] env[61867]: DEBUG nova.network.neutron [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.312269] env[61867]: DEBUG nova.network.neutron [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.365209] env[61867]: DEBUG oslo_concurrency.lockutils [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Acquiring lock "7c0badee-0e2b-4a13-89b1-e57df5c2c9ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.365460] env[61867]: DEBUG oslo_concurrency.lockutils [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Lock "7c0badee-0e2b-4a13-89b1-e57df5c2c9ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.550687] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Releasing lock "refresh_cache-45e2004d-731b-4002-84b2-3f9138c09663" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.550955] env[61867]: DEBUG nova.compute.manager [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 615.551114] env[61867]: DEBUG nova.compute.manager [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.551274] env[61867]: DEBUG nova.network.neutron [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 615.567880] env[61867]: DEBUG nova.network.neutron [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.804234] env[61867]: DEBUG nova.compute.manager [req-3fe065fc-2cea-48e2-ad02-f34800c15750 req-fa7c5334-f3bd-41c8-9adc-fd732293c997 service nova] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Received event network-changed-6d020051-3a6c-4f84-922b-85cfee5b7411 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 615.804434] env[61867]: DEBUG nova.compute.manager [req-3fe065fc-2cea-48e2-ad02-f34800c15750 req-fa7c5334-f3bd-41c8-9adc-fd732293c997 service nova] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Refreshing instance network info cache due to event network-changed-6d020051-3a6c-4f84-922b-85cfee5b7411. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 615.804686] env[61867]: DEBUG oslo_concurrency.lockutils [req-3fe065fc-2cea-48e2-ad02-f34800c15750 req-fa7c5334-f3bd-41c8-9adc-fd732293c997 service nova] Acquiring lock "refresh_cache-ba4cc401-7ae3-40bc-996c-13741a65d991" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.814368] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Releasing lock "refresh_cache-ba4cc401-7ae3-40bc-996c-13741a65d991" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.814861] env[61867]: DEBUG nova.compute.manager [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 615.815096] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 615.815354] env[61867]: DEBUG oslo_concurrency.lockutils [req-3fe065fc-2cea-48e2-ad02-f34800c15750 req-fa7c5334-f3bd-41c8-9adc-fd732293c997 service nova] Acquired lock "refresh_cache-ba4cc401-7ae3-40bc-996c-13741a65d991" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.815522] env[61867]: DEBUG nova.network.neutron [req-3fe065fc-2cea-48e2-ad02-f34800c15750 req-fa7c5334-f3bd-41c8-9adc-fd732293c997 service nova] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Refreshing network info cache for port 6d020051-3a6c-4f84-922b-85cfee5b7411 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 615.816474] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-82dd4b3d-2e30-4dd2-849d-af10d3a0303a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.822924] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0db3d4f2-8258-4d13-935e-f3a9bf264e77 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.829739] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30ddfce9-9312-40c6-892d-985641c828e1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.843167] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7555eeb9-c055-49bb-a3f7-ffd78ecf45d5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.853267] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ba4cc401-7ae3-40bc-996c-13741a65d991 could not be found. [ 615.853267] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 615.853413] env[61867]: INFO nova.compute.manager [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Took 0.04 seconds to destroy the instance on the hypervisor. [ 615.853848] env[61867]: DEBUG oslo.service.loopingcall [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 615.878685] env[61867]: DEBUG nova.compute.manager [-] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.878814] env[61867]: DEBUG nova.network.neutron [-] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 615.881724] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb0ce5c5-22b4-4cc2-9c41-e98f41f0ca23 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.889182] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb2906c-82ba-4a53-881c-a9d1b4049d26 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.903674] env[61867]: DEBUG nova.compute.provider_tree [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.905425] env[61867]: DEBUG nova.network.neutron [-] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.070837] env[61867]: DEBUG nova.network.neutron [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.335943] env[61867]: DEBUG nova.network.neutron [req-3fe065fc-2cea-48e2-ad02-f34800c15750 req-fa7c5334-f3bd-41c8-9adc-fd732293c997 service nova] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.408024] env[61867]: DEBUG nova.scheduler.client.report [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.411250] env[61867]: DEBUG nova.network.neutron [-] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.413110] env[61867]: DEBUG nova.network.neutron [req-3fe065fc-2cea-48e2-ad02-f34800c15750 req-fa7c5334-f3bd-41c8-9adc-fd732293c997 service nova] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.573354] env[61867]: INFO nova.compute.manager [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] [instance: 45e2004d-731b-4002-84b2-3f9138c09663] Took 1.02 seconds to deallocate network for instance. [ 616.915137] env[61867]: DEBUG oslo_concurrency.lockutils [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.478s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.915719] env[61867]: DEBUG nova.compute.manager [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 616.919027] env[61867]: INFO nova.compute.manager [-] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Took 1.04 seconds to deallocate network for instance. [ 616.919175] env[61867]: DEBUG oslo_concurrency.lockutils [req-3fe065fc-2cea-48e2-ad02-f34800c15750 req-fa7c5334-f3bd-41c8-9adc-fd732293c997 service nova] Releasing lock "refresh_cache-ba4cc401-7ae3-40bc-996c-13741a65d991" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.919494] env[61867]: DEBUG nova.compute.manager [req-3fe065fc-2cea-48e2-ad02-f34800c15750 req-fa7c5334-f3bd-41c8-9adc-fd732293c997 service nova] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Received event network-vif-deleted-6d020051-3a6c-4f84-922b-85cfee5b7411 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 616.920286] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.520s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.920548] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.920727] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61867) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 616.921011] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.819s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.922429] env[61867]: INFO nova.compute.claims [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 616.926766] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20f6ad0a-04ab-4952-8472-b1c38f6ca671 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.929653] env[61867]: DEBUG nova.compute.claims [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 616.929853] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.935291] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a977fdd0-d1f9-42dc-bb48-a0e7ce44c92e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.949922] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae13e8a5-43eb-4518-88fc-ed1991cc3b90 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.957028] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd1b2ad-2859-4fc1-aba3-290c7d795a67 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.987216] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181483MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61867) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 616.987374] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.422237] env[61867]: DEBUG nova.compute.utils [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 617.423666] env[61867]: DEBUG nova.compute.manager [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 617.423845] env[61867]: DEBUG nova.network.neutron [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 617.462675] env[61867]: DEBUG nova.policy [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a03f16307d2a401bbd8d74c579440aa7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6d4db4879810418fa26e7a1e5ef6a350', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 617.610382] env[61867]: INFO nova.scheduler.client.report [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Deleted allocations for instance 45e2004d-731b-4002-84b2-3f9138c09663 [ 617.748854] env[61867]: DEBUG nova.network.neutron [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Successfully created port: cdec95b8-f286-44cd-8d19-77aeac2b74fa {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 617.927494] env[61867]: DEBUG nova.compute.manager [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 618.118351] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54675f94-5e36-4f1a-a9b3-66431789da1b tempest-ListServerFiltersTestJSON-579140893 tempest-ListServerFiltersTestJSON-579140893-project-member] Lock "45e2004d-731b-4002-84b2-3f9138c09663" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.987s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.349019] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd500fe-165e-4cc2-9f91-e7cec9008815 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.354049] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c0f23ac-3748-4813-9ec6-b803ee4c5f6d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.385558] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace91c5d-0f2d-4d77-a455-91157c043d10 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.393121] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c982aa9b-1270-4632-a7ce-33cc85b1a8c5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.406396] env[61867]: DEBUG nova.compute.provider_tree [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.578837] env[61867]: DEBUG nova.compute.manager [req-61b09a83-7772-4677-af7f-f7b76ad783c6 req-f53ca29b-e430-460a-9a92-daa3c6fec5a6 service nova] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Received event network-changed-cdec95b8-f286-44cd-8d19-77aeac2b74fa {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 618.579174] env[61867]: DEBUG nova.compute.manager [req-61b09a83-7772-4677-af7f-f7b76ad783c6 req-f53ca29b-e430-460a-9a92-daa3c6fec5a6 service nova] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Refreshing instance network info cache due to event network-changed-cdec95b8-f286-44cd-8d19-77aeac2b74fa. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 618.579452] env[61867]: DEBUG oslo_concurrency.lockutils [req-61b09a83-7772-4677-af7f-f7b76ad783c6 req-f53ca29b-e430-460a-9a92-daa3c6fec5a6 service nova] Acquiring lock "refresh_cache-013fc4b2-95f7-439e-b82b-da2265ccf736" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.579668] env[61867]: DEBUG oslo_concurrency.lockutils [req-61b09a83-7772-4677-af7f-f7b76ad783c6 req-f53ca29b-e430-460a-9a92-daa3c6fec5a6 service nova] Acquired lock "refresh_cache-013fc4b2-95f7-439e-b82b-da2265ccf736" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.579865] env[61867]: DEBUG nova.network.neutron [req-61b09a83-7772-4677-af7f-f7b76ad783c6 req-f53ca29b-e430-460a-9a92-daa3c6fec5a6 service nova] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Refreshing network info cache for port cdec95b8-f286-44cd-8d19-77aeac2b74fa {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 618.623758] env[61867]: DEBUG nova.compute.manager [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 618.703203] env[61867]: ERROR nova.compute.manager [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cdec95b8-f286-44cd-8d19-77aeac2b74fa, please check neutron logs for more information. [ 618.703203] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 618.703203] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.703203] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 618.703203] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.703203] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 618.703203] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.703203] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 618.703203] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.703203] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 618.703203] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.703203] env[61867]: ERROR nova.compute.manager raise self.value [ 618.703203] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.703203] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 618.703203] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.703203] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 618.703624] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.703624] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 618.703624] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cdec95b8-f286-44cd-8d19-77aeac2b74fa, please check neutron logs for more information. [ 618.703624] env[61867]: ERROR nova.compute.manager [ 618.703624] env[61867]: Traceback (most recent call last): [ 618.703624] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 618.703624] env[61867]: listener.cb(fileno) [ 618.703624] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.703624] env[61867]: result = function(*args, **kwargs) [ 618.703624] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.703624] env[61867]: return func(*args, **kwargs) [ 618.703624] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.703624] env[61867]: raise e [ 618.703624] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.703624] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 618.703624] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.703624] env[61867]: created_port_ids = self._update_ports_for_instance( [ 618.703624] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.703624] env[61867]: with excutils.save_and_reraise_exception(): [ 618.703624] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.703624] env[61867]: self.force_reraise() [ 618.703624] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.703624] env[61867]: raise self.value [ 618.703624] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.703624] env[61867]: updated_port = self._update_port( [ 618.703624] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.703624] env[61867]: _ensure_no_port_binding_failure(port) [ 618.703624] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.703624] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 618.704406] env[61867]: nova.exception.PortBindingFailed: Binding failed for port cdec95b8-f286-44cd-8d19-77aeac2b74fa, please check neutron logs for more information. [ 618.704406] env[61867]: Removing descriptor: 21 [ 618.910184] env[61867]: DEBUG nova.scheduler.client.report [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.939182] env[61867]: DEBUG nova.compute.manager [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 618.963936] env[61867]: DEBUG nova.virt.hardware [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 618.964203] env[61867]: DEBUG nova.virt.hardware [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 618.964357] env[61867]: DEBUG nova.virt.hardware [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 618.964533] env[61867]: DEBUG nova.virt.hardware [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 618.964676] env[61867]: DEBUG nova.virt.hardware [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 618.964819] env[61867]: DEBUG nova.virt.hardware [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 618.965033] env[61867]: DEBUG nova.virt.hardware [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 618.965193] env[61867]: DEBUG nova.virt.hardware [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 618.965359] env[61867]: DEBUG nova.virt.hardware [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 618.965519] env[61867]: DEBUG nova.virt.hardware [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 618.965691] env[61867]: DEBUG nova.virt.hardware [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.966832] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc725795-6e48-4a75-a43e-6d385b819c35 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.975525] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b9c7186-ae6d-4772-8411-b617fb9c4962 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.990629] env[61867]: ERROR nova.compute.manager [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cdec95b8-f286-44cd-8d19-77aeac2b74fa, please check neutron logs for more information. [ 618.990629] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Traceback (most recent call last): [ 618.990629] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 618.990629] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] yield resources [ 618.990629] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.990629] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] self.driver.spawn(context, instance, image_meta, [ 618.990629] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 618.990629] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.990629] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.990629] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] vm_ref = self.build_virtual_machine(instance, [ 618.990629] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.990985] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.990985] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.990985] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] for vif in network_info: [ 618.990985] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.990985] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] return self._sync_wrapper(fn, *args, **kwargs) [ 618.990985] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.990985] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] self.wait() [ 618.990985] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.990985] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] self[:] = self._gt.wait() [ 618.990985] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.990985] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] return self._exit_event.wait() [ 618.990985] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 618.990985] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] current.throw(*self._exc) [ 618.991385] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.991385] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] result = function(*args, **kwargs) [ 618.991385] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.991385] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] return func(*args, **kwargs) [ 618.991385] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.991385] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] raise e [ 618.991385] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.991385] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] nwinfo = self.network_api.allocate_for_instance( [ 618.991385] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.991385] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] created_port_ids = self._update_ports_for_instance( [ 618.991385] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.991385] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] with excutils.save_and_reraise_exception(): [ 618.991385] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.991705] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] self.force_reraise() [ 618.991705] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.991705] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] raise self.value [ 618.991705] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.991705] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] updated_port = self._update_port( [ 618.991705] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.991705] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] _ensure_no_port_binding_failure(port) [ 618.991705] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.991705] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] raise exception.PortBindingFailed(port_id=port['id']) [ 618.991705] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] nova.exception.PortBindingFailed: Binding failed for port cdec95b8-f286-44cd-8d19-77aeac2b74fa, please check neutron logs for more information. [ 618.991705] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] [ 618.991705] env[61867]: INFO nova.compute.manager [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Terminating instance [ 618.994540] env[61867]: DEBUG oslo_concurrency.lockutils [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Acquiring lock "refresh_cache-013fc4b2-95f7-439e-b82b-da2265ccf736" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.098531] env[61867]: DEBUG nova.network.neutron [req-61b09a83-7772-4677-af7f-f7b76ad783c6 req-f53ca29b-e430-460a-9a92-daa3c6fec5a6 service nova] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.144188] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.177998] env[61867]: DEBUG nova.network.neutron [req-61b09a83-7772-4677-af7f-f7b76ad783c6 req-f53ca29b-e430-460a-9a92-daa3c6fec5a6 service nova] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.415440] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.416446] env[61867]: DEBUG nova.compute.manager [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 619.419449] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.261s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.682056] env[61867]: DEBUG oslo_concurrency.lockutils [req-61b09a83-7772-4677-af7f-f7b76ad783c6 req-f53ca29b-e430-460a-9a92-daa3c6fec5a6 service nova] Releasing lock "refresh_cache-013fc4b2-95f7-439e-b82b-da2265ccf736" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.682498] env[61867]: DEBUG oslo_concurrency.lockutils [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Acquired lock "refresh_cache-013fc4b2-95f7-439e-b82b-da2265ccf736" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.682690] env[61867]: DEBUG nova.network.neutron [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.925488] env[61867]: DEBUG nova.compute.utils [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 619.927197] env[61867]: DEBUG nova.compute.manager [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 619.927346] env[61867]: DEBUG nova.network.neutron [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 619.991205] env[61867]: DEBUG nova.policy [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2c1a78d5f8c41eaa01e3a0ccfd259ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78c08ac938a643f0b74f2050e4b8d2c3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 620.209331] env[61867]: DEBUG nova.network.neutron [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.331938] env[61867]: DEBUG nova.network.neutron [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.343831] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4541fe-1a38-41bd-ae75-06d9709ed232 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.352576] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c4eeeb-a773-4cf9-b68f-93d58649843b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.386933] env[61867]: DEBUG nova.network.neutron [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Successfully created port: a83e557f-96a5-4bde-83ce-d197aece2270 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 620.389436] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1988148-7cba-462e-b412-66436a362ae2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.396657] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3936e9d4-d2d6-4ce5-ba60-8e0a134038f3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.410258] env[61867]: DEBUG nova.compute.provider_tree [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.433369] env[61867]: DEBUG nova.compute.manager [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 620.705677] env[61867]: DEBUG nova.compute.manager [req-bd63cf15-6f81-49ad-b3e1-571f4dc165e4 req-4de13604-04e3-4070-ad47-b768f55a7c5c service nova] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Received event network-vif-deleted-cdec95b8-f286-44cd-8d19-77aeac2b74fa {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 620.834357] env[61867]: DEBUG oslo_concurrency.lockutils [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Releasing lock "refresh_cache-013fc4b2-95f7-439e-b82b-da2265ccf736" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.834813] env[61867]: DEBUG nova.compute.manager [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 620.834997] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 620.835686] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bc68d824-233b-486b-a970-2e988fb13b23 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.844131] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe34c3f-3798-40ea-b861-7c6fe7f26196 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.867553] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 013fc4b2-95f7-439e-b82b-da2265ccf736 could not be found. [ 620.867796] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 620.867983] env[61867]: INFO nova.compute.manager [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Took 0.03 seconds to destroy the instance on the hypervisor. [ 620.868271] env[61867]: DEBUG oslo.service.loopingcall [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 620.868495] env[61867]: DEBUG nova.compute.manager [-] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.868587] env[61867]: DEBUG nova.network.neutron [-] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.884145] env[61867]: DEBUG nova.network.neutron [-] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.912883] env[61867]: DEBUG nova.scheduler.client.report [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.087391] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Acquiring lock "d0db58a5-d773-4e45-a238-07cfc156bdd2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.087629] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Lock "d0db58a5-d773-4e45-a238-07cfc156bdd2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.258999] env[61867]: ERROR nova.compute.manager [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a83e557f-96a5-4bde-83ce-d197aece2270, please check neutron logs for more information. [ 621.258999] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 621.258999] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.258999] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 621.258999] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.258999] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 621.258999] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.258999] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 621.258999] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.258999] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 621.258999] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.258999] env[61867]: ERROR nova.compute.manager raise self.value [ 621.258999] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.258999] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 621.258999] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.258999] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 621.259545] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.259545] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 621.259545] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a83e557f-96a5-4bde-83ce-d197aece2270, please check neutron logs for more information. [ 621.259545] env[61867]: ERROR nova.compute.manager [ 621.259545] env[61867]: Traceback (most recent call last): [ 621.259545] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 621.259545] env[61867]: listener.cb(fileno) [ 621.259545] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.259545] env[61867]: result = function(*args, **kwargs) [ 621.259545] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 621.259545] env[61867]: return func(*args, **kwargs) [ 621.259545] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.259545] env[61867]: raise e [ 621.259545] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.259545] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 621.259545] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.259545] env[61867]: created_port_ids = self._update_ports_for_instance( [ 621.259545] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.259545] env[61867]: with excutils.save_and_reraise_exception(): [ 621.259545] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.259545] env[61867]: self.force_reraise() [ 621.259545] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.259545] env[61867]: raise self.value [ 621.259545] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.259545] env[61867]: updated_port = self._update_port( [ 621.259545] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.259545] env[61867]: _ensure_no_port_binding_failure(port) [ 621.259545] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.259545] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 621.260684] env[61867]: nova.exception.PortBindingFailed: Binding failed for port a83e557f-96a5-4bde-83ce-d197aece2270, please check neutron logs for more information. [ 621.260684] env[61867]: Removing descriptor: 21 [ 621.388103] env[61867]: DEBUG nova.network.neutron [-] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.417660] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.998s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.418361] env[61867]: ERROR nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3ca41fad-6498-4d3f-ac59-22f2e90bcbbe, please check neutron logs for more information. [ 621.418361] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] Traceback (most recent call last): [ 621.418361] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.418361] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] self.driver.spawn(context, instance, image_meta, [ 621.418361] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 621.418361] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.418361] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.418361] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] vm_ref = self.build_virtual_machine(instance, [ 621.418361] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.418361] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.418361] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.418668] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] for vif in network_info: [ 621.418668] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 621.418668] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] return self._sync_wrapper(fn, *args, **kwargs) [ 621.418668] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 621.418668] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] self.wait() [ 621.418668] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 621.418668] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] self[:] = self._gt.wait() [ 621.418668] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.418668] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] return self._exit_event.wait() [ 621.418668] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 621.418668] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] result = hub.switch() [ 621.418668] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 621.418668] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] return self.greenlet.switch() [ 621.418989] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.418989] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] result = function(*args, **kwargs) [ 621.418989] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 621.418989] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] return func(*args, **kwargs) [ 621.418989] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.418989] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] raise e [ 621.418989] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.418989] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] nwinfo = self.network_api.allocate_for_instance( [ 621.418989] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.418989] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] created_port_ids = self._update_ports_for_instance( [ 621.418989] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.418989] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] with excutils.save_and_reraise_exception(): [ 621.418989] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.419615] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] self.force_reraise() [ 621.419615] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.419615] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] raise self.value [ 621.419615] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.419615] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] updated_port = self._update_port( [ 621.419615] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.419615] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] _ensure_no_port_binding_failure(port) [ 621.419615] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.419615] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] raise exception.PortBindingFailed(port_id=port['id']) [ 621.419615] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] nova.exception.PortBindingFailed: Binding failed for port 3ca41fad-6498-4d3f-ac59-22f2e90bcbbe, please check neutron logs for more information. [ 621.419615] env[61867]: ERROR nova.compute.manager [instance: 2ae534f8-1921-4892-8556-e2732624f412] [ 621.420059] env[61867]: DEBUG nova.compute.utils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Binding failed for port 3ca41fad-6498-4d3f-ac59-22f2e90bcbbe, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 621.420571] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.099s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.423593] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Build of instance 2ae534f8-1921-4892-8556-e2732624f412 was re-scheduled: Binding failed for port 3ca41fad-6498-4d3f-ac59-22f2e90bcbbe, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 621.423727] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 621.423971] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "refresh_cache-2ae534f8-1921-4892-8556-e2732624f412" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.424145] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquired lock "refresh_cache-2ae534f8-1921-4892-8556-e2732624f412" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.424309] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.444028] env[61867]: DEBUG nova.compute.manager [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 621.467135] env[61867]: DEBUG nova.virt.hardware [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 621.467390] env[61867]: DEBUG nova.virt.hardware [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 621.467546] env[61867]: DEBUG nova.virt.hardware [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 621.467775] env[61867]: DEBUG nova.virt.hardware [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 621.467927] env[61867]: DEBUG nova.virt.hardware [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 621.468088] env[61867]: DEBUG nova.virt.hardware [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 621.468297] env[61867]: DEBUG nova.virt.hardware [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 621.468454] env[61867]: DEBUG nova.virt.hardware [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 621.468616] env[61867]: DEBUG nova.virt.hardware [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 621.468778] env[61867]: DEBUG nova.virt.hardware [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 621.468948] env[61867]: DEBUG nova.virt.hardware [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 621.469893] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e89f68-f696-4199-89a0-af3a2bb053bb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.477621] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06902143-a305-4ac7-8398-5287938e5964 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.490833] env[61867]: ERROR nova.compute.manager [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a83e557f-96a5-4bde-83ce-d197aece2270, please check neutron logs for more information. [ 621.490833] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Traceback (most recent call last): [ 621.490833] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 621.490833] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] yield resources [ 621.490833] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.490833] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] self.driver.spawn(context, instance, image_meta, [ 621.490833] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 621.490833] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.490833] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.490833] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] vm_ref = self.build_virtual_machine(instance, [ 621.490833] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.491208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.491208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.491208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] for vif in network_info: [ 621.491208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 621.491208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] return self._sync_wrapper(fn, *args, **kwargs) [ 621.491208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 621.491208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] self.wait() [ 621.491208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 621.491208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] self[:] = self._gt.wait() [ 621.491208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.491208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] return self._exit_event.wait() [ 621.491208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 621.491208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] current.throw(*self._exc) [ 621.491576] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.491576] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] result = function(*args, **kwargs) [ 621.491576] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 621.491576] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] return func(*args, **kwargs) [ 621.491576] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.491576] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] raise e [ 621.491576] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.491576] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] nwinfo = self.network_api.allocate_for_instance( [ 621.491576] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.491576] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] created_port_ids = self._update_ports_for_instance( [ 621.491576] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.491576] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] with excutils.save_and_reraise_exception(): [ 621.491576] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.491939] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] self.force_reraise() [ 621.491939] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.491939] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] raise self.value [ 621.491939] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.491939] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] updated_port = self._update_port( [ 621.491939] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.491939] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] _ensure_no_port_binding_failure(port) [ 621.491939] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.491939] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] raise exception.PortBindingFailed(port_id=port['id']) [ 621.491939] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] nova.exception.PortBindingFailed: Binding failed for port a83e557f-96a5-4bde-83ce-d197aece2270, please check neutron logs for more information. [ 621.491939] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] [ 621.491939] env[61867]: INFO nova.compute.manager [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Terminating instance [ 621.492890] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Acquiring lock "refresh_cache-32e947ce-c7e5-4568-b1e4-89dfc6a8009e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.493055] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Acquired lock "refresh_cache-32e947ce-c7e5-4568-b1e4-89dfc6a8009e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.493224] env[61867]: DEBUG nova.network.neutron [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.890797] env[61867]: INFO nova.compute.manager [-] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Took 1.02 seconds to deallocate network for instance. [ 621.893122] env[61867]: DEBUG nova.compute.claims [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 621.893360] env[61867]: DEBUG oslo_concurrency.lockutils [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.945732] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.016502] env[61867]: DEBUG nova.network.neutron [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.065543] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.084242] env[61867]: DEBUG nova.network.neutron [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.304674] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc0773a-7475-4172-9e14-65ca56b4cd4f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.311905] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5360ee9d-c1ea-4f1e-85f0-09b5a8c6019c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.340102] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dca5604-bd81-4719-8eeb-bf7099141c7f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.347022] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bbba65a-2b43-473d-9388-d7d980e1dee8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.359842] env[61867]: DEBUG nova.compute.provider_tree [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.568211] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Releasing lock "refresh_cache-2ae534f8-1921-4892-8556-e2732624f412" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.568452] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 622.568672] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.568804] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.584259] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.589693] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Releasing lock "refresh_cache-32e947ce-c7e5-4568-b1e4-89dfc6a8009e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.589693] env[61867]: DEBUG nova.compute.manager [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 622.589861] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 622.590215] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ba2276f3-7e09-432a-98da-8e148691594f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.599955] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50856d0-97e7-41fe-a139-7d48bf87a536 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.621794] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 32e947ce-c7e5-4568-b1e4-89dfc6a8009e could not be found. [ 622.621986] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 622.622183] env[61867]: INFO nova.compute.manager [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 622.622417] env[61867]: DEBUG oslo.service.loopingcall [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 622.622613] env[61867]: DEBUG nova.compute.manager [-] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.622703] env[61867]: DEBUG nova.network.neutron [-] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.636537] env[61867]: DEBUG nova.network.neutron [-] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.749046] env[61867]: DEBUG nova.compute.manager [req-5c2ef266-dace-4437-af3c-af0d43e35520 req-149be3f5-f42f-40cc-ab3b-5326a030c17c service nova] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Received event network-changed-a83e557f-96a5-4bde-83ce-d197aece2270 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 622.749251] env[61867]: DEBUG nova.compute.manager [req-5c2ef266-dace-4437-af3c-af0d43e35520 req-149be3f5-f42f-40cc-ab3b-5326a030c17c service nova] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Refreshing instance network info cache due to event network-changed-a83e557f-96a5-4bde-83ce-d197aece2270. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 622.749485] env[61867]: DEBUG oslo_concurrency.lockutils [req-5c2ef266-dace-4437-af3c-af0d43e35520 req-149be3f5-f42f-40cc-ab3b-5326a030c17c service nova] Acquiring lock "refresh_cache-32e947ce-c7e5-4568-b1e4-89dfc6a8009e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.749649] env[61867]: DEBUG oslo_concurrency.lockutils [req-5c2ef266-dace-4437-af3c-af0d43e35520 req-149be3f5-f42f-40cc-ab3b-5326a030c17c service nova] Acquired lock "refresh_cache-32e947ce-c7e5-4568-b1e4-89dfc6a8009e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.749810] env[61867]: DEBUG nova.network.neutron [req-5c2ef266-dace-4437-af3c-af0d43e35520 req-149be3f5-f42f-40cc-ab3b-5326a030c17c service nova] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Refreshing network info cache for port a83e557f-96a5-4bde-83ce-d197aece2270 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 622.863942] env[61867]: DEBUG nova.scheduler.client.report [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.086551] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.138382] env[61867]: DEBUG nova.network.neutron [-] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.267924] env[61867]: DEBUG nova.network.neutron [req-5c2ef266-dace-4437-af3c-af0d43e35520 req-149be3f5-f42f-40cc-ab3b-5326a030c17c service nova] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.311799] env[61867]: DEBUG nova.network.neutron [req-5c2ef266-dace-4437-af3c-af0d43e35520 req-149be3f5-f42f-40cc-ab3b-5326a030c17c service nova] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.369288] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.949s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.370498] env[61867]: ERROR nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8c5f6b14-4a4d-49f9-a67a-99aadc4b4211, please check neutron logs for more information. [ 623.370498] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Traceback (most recent call last): [ 623.370498] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 623.370498] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] self.driver.spawn(context, instance, image_meta, [ 623.370498] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 623.370498] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.370498] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.370498] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] vm_ref = self.build_virtual_machine(instance, [ 623.370498] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.370498] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.370498] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.370927] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] for vif in network_info: [ 623.370927] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 623.370927] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] return self._sync_wrapper(fn, *args, **kwargs) [ 623.370927] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 623.370927] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] self.wait() [ 623.370927] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 623.370927] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] self[:] = self._gt.wait() [ 623.370927] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.370927] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] return self._exit_event.wait() [ 623.370927] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.370927] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] result = hub.switch() [ 623.370927] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.370927] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] return self.greenlet.switch() [ 623.371398] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.371398] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] result = function(*args, **kwargs) [ 623.371398] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 623.371398] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] return func(*args, **kwargs) [ 623.371398] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.371398] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] raise e [ 623.371398] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.371398] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] nwinfo = self.network_api.allocate_for_instance( [ 623.371398] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.371398] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] created_port_ids = self._update_ports_for_instance( [ 623.371398] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.371398] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] with excutils.save_and_reraise_exception(): [ 623.371398] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.371856] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] self.force_reraise() [ 623.371856] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.371856] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] raise self.value [ 623.371856] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.371856] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] updated_port = self._update_port( [ 623.371856] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.371856] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] _ensure_no_port_binding_failure(port) [ 623.371856] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.371856] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] raise exception.PortBindingFailed(port_id=port['id']) [ 623.371856] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] nova.exception.PortBindingFailed: Binding failed for port 8c5f6b14-4a4d-49f9-a67a-99aadc4b4211, please check neutron logs for more information. [ 623.371856] env[61867]: ERROR nova.compute.manager [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] [ 623.372256] env[61867]: DEBUG nova.compute.utils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Binding failed for port 8c5f6b14-4a4d-49f9-a67a-99aadc4b4211, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 623.372256] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.881s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.373831] env[61867]: INFO nova.compute.claims [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 623.377119] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Build of instance 18b98634-d76d-496b-904f-1a9c289cdb8a was re-scheduled: Binding failed for port 8c5f6b14-4a4d-49f9-a67a-99aadc4b4211, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 623.377556] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 623.377918] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "refresh_cache-18b98634-d76d-496b-904f-1a9c289cdb8a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.378143] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquired lock "refresh_cache-18b98634-d76d-496b-904f-1a9c289cdb8a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.378313] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 623.589318] env[61867]: INFO nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 2ae534f8-1921-4892-8556-e2732624f412] Took 1.02 seconds to deallocate network for instance. [ 623.641444] env[61867]: INFO nova.compute.manager [-] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Took 1.02 seconds to deallocate network for instance. [ 623.643514] env[61867]: DEBUG nova.compute.claims [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 623.643707] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.814384] env[61867]: DEBUG oslo_concurrency.lockutils [req-5c2ef266-dace-4437-af3c-af0d43e35520 req-149be3f5-f42f-40cc-ab3b-5326a030c17c service nova] Releasing lock "refresh_cache-32e947ce-c7e5-4568-b1e4-89dfc6a8009e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.814384] env[61867]: DEBUG nova.compute.manager [req-5c2ef266-dace-4437-af3c-af0d43e35520 req-149be3f5-f42f-40cc-ab3b-5326a030c17c service nova] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Received event network-vif-deleted-a83e557f-96a5-4bde-83ce-d197aece2270 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 623.897024] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.967113] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.469204] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Releasing lock "refresh_cache-18b98634-d76d-496b-904f-1a9c289cdb8a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.469459] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 624.469627] env[61867]: DEBUG nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 624.469788] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 624.486717] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.612528] env[61867]: INFO nova.scheduler.client.report [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Deleted allocations for instance 2ae534f8-1921-4892-8556-e2732624f412 [ 624.764012] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-291dabe2-d395-47f2-b389-8bbac2577131 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.771816] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec11efe-f7a2-413d-97dd-464e012305b0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.805331] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d114459-0c58-4b99-9801-a5f6f58b17f3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.815183] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ef75c9-2463-4bca-b09f-80d6e50ec1e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.831955] env[61867]: DEBUG nova.compute.provider_tree [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.989755] env[61867]: DEBUG nova.network.neutron [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.121496] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "2ae534f8-1921-4892-8556-e2732624f412" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.314s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.334510] env[61867]: DEBUG nova.scheduler.client.report [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 625.493096] env[61867]: INFO nova.compute.manager [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 18b98634-d76d-496b-904f-1a9c289cdb8a] Took 1.02 seconds to deallocate network for instance. [ 625.626568] env[61867]: DEBUG nova.compute.manager [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.839806] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.468s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.840390] env[61867]: DEBUG nova.compute.manager [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 625.843167] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.430s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.148918] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.348491] env[61867]: DEBUG nova.compute.utils [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 626.353026] env[61867]: DEBUG nova.compute.manager [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 626.353026] env[61867]: DEBUG nova.network.neutron [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 626.404684] env[61867]: DEBUG nova.policy [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1a9fd65b1fe34eebbd63c69a0359567c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0aee94970ab6413fb7188d5692bfd116', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 626.525016] env[61867]: INFO nova.scheduler.client.report [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Deleted allocations for instance 18b98634-d76d-496b-904f-1a9c289cdb8a [ 626.660753] env[61867]: DEBUG nova.network.neutron [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Successfully created port: 4f1af8bd-6c9c-433f-8671-5beba8b94491 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 626.730879] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27275eec-79a7-4e49-923c-760a273e3cbc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.738486] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61427d43-2e27-4259-958b-a213de56f910 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.769119] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11df084f-6877-4f16-bc4a-ec54ecc5c07c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.776039] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa52cc7-d950-4f84-9a4e-9905bcc08f44 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.788950] env[61867]: DEBUG nova.compute.provider_tree [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.858608] env[61867]: DEBUG nova.compute.manager [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 627.033807] env[61867]: DEBUG oslo_concurrency.lockutils [None req-709da486-cc50-4689-85b9-2773e3fbf1f5 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "18b98634-d76d-496b-904f-1a9c289cdb8a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.200s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.293050] env[61867]: DEBUG nova.scheduler.client.report [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.449191] env[61867]: DEBUG nova.compute.manager [req-37aa9944-2f3c-4609-8e84-e21e6bd42e9f req-ad2a82b5-c421-41b4-a67a-5c29ffe4a464 service nova] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Received event network-changed-4f1af8bd-6c9c-433f-8671-5beba8b94491 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 627.449562] env[61867]: DEBUG nova.compute.manager [req-37aa9944-2f3c-4609-8e84-e21e6bd42e9f req-ad2a82b5-c421-41b4-a67a-5c29ffe4a464 service nova] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Refreshing instance network info cache due to event network-changed-4f1af8bd-6c9c-433f-8671-5beba8b94491. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 627.449662] env[61867]: DEBUG oslo_concurrency.lockutils [req-37aa9944-2f3c-4609-8e84-e21e6bd42e9f req-ad2a82b5-c421-41b4-a67a-5c29ffe4a464 service nova] Acquiring lock "refresh_cache-9ffe4fa1-4239-435b-ac7e-648cec92af65" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.449775] env[61867]: DEBUG oslo_concurrency.lockutils [req-37aa9944-2f3c-4609-8e84-e21e6bd42e9f req-ad2a82b5-c421-41b4-a67a-5c29ffe4a464 service nova] Acquired lock "refresh_cache-9ffe4fa1-4239-435b-ac7e-648cec92af65" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.449896] env[61867]: DEBUG nova.network.neutron [req-37aa9944-2f3c-4609-8e84-e21e6bd42e9f req-ad2a82b5-c421-41b4-a67a-5c29ffe4a464 service nova] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Refreshing network info cache for port 4f1af8bd-6c9c-433f-8671-5beba8b94491 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 627.493434] env[61867]: ERROR nova.compute.manager [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4f1af8bd-6c9c-433f-8671-5beba8b94491, please check neutron logs for more information. [ 627.493434] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 627.493434] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.493434] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 627.493434] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.493434] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 627.493434] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.493434] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 627.493434] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.493434] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 627.493434] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.493434] env[61867]: ERROR nova.compute.manager raise self.value [ 627.493434] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.493434] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 627.493434] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.493434] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 627.493882] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.493882] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 627.493882] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4f1af8bd-6c9c-433f-8671-5beba8b94491, please check neutron logs for more information. [ 627.493882] env[61867]: ERROR nova.compute.manager [ 627.493882] env[61867]: Traceback (most recent call last): [ 627.493882] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 627.493882] env[61867]: listener.cb(fileno) [ 627.493882] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.493882] env[61867]: result = function(*args, **kwargs) [ 627.493882] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.493882] env[61867]: return func(*args, **kwargs) [ 627.493882] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.493882] env[61867]: raise e [ 627.493882] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.493882] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 627.493882] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.493882] env[61867]: created_port_ids = self._update_ports_for_instance( [ 627.493882] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.493882] env[61867]: with excutils.save_and_reraise_exception(): [ 627.493882] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.493882] env[61867]: self.force_reraise() [ 627.493882] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.493882] env[61867]: raise self.value [ 627.493882] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.493882] env[61867]: updated_port = self._update_port( [ 627.493882] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.493882] env[61867]: _ensure_no_port_binding_failure(port) [ 627.493882] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.493882] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 627.494701] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 4f1af8bd-6c9c-433f-8671-5beba8b94491, please check neutron logs for more information. [ 627.494701] env[61867]: Removing descriptor: 17 [ 627.538621] env[61867]: DEBUG nova.compute.manager [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 627.798815] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.956s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.799740] env[61867]: ERROR nova.compute.manager [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 06047bd3-a41f-45f1-9e3f-7b129fd2a39a, please check neutron logs for more information. [ 627.799740] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Traceback (most recent call last): [ 627.799740] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 627.799740] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] self.driver.spawn(context, instance, image_meta, [ 627.799740] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 627.799740] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.799740] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.799740] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] vm_ref = self.build_virtual_machine(instance, [ 627.799740] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.799740] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.799740] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.800229] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] for vif in network_info: [ 627.800229] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.800229] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] return self._sync_wrapper(fn, *args, **kwargs) [ 627.800229] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.800229] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] self.wait() [ 627.800229] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.800229] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] self[:] = self._gt.wait() [ 627.800229] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.800229] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] return self._exit_event.wait() [ 627.800229] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 627.800229] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] current.throw(*self._exc) [ 627.800229] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.800229] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] result = function(*args, **kwargs) [ 627.800594] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.800594] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] return func(*args, **kwargs) [ 627.800594] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.800594] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] raise e [ 627.800594] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.800594] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] nwinfo = self.network_api.allocate_for_instance( [ 627.800594] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.800594] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] created_port_ids = self._update_ports_for_instance( [ 627.800594] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.800594] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] with excutils.save_and_reraise_exception(): [ 627.800594] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.800594] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] self.force_reraise() [ 627.800594] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.801010] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] raise self.value [ 627.801010] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.801010] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] updated_port = self._update_port( [ 627.801010] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.801010] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] _ensure_no_port_binding_failure(port) [ 627.801010] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.801010] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] raise exception.PortBindingFailed(port_id=port['id']) [ 627.801010] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] nova.exception.PortBindingFailed: Binding failed for port 06047bd3-a41f-45f1-9e3f-7b129fd2a39a, please check neutron logs for more information. [ 627.801010] env[61867]: ERROR nova.compute.manager [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] [ 627.801010] env[61867]: DEBUG nova.compute.utils [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Binding failed for port 06047bd3-a41f-45f1-9e3f-7b129fd2a39a, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 627.801865] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.764s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.803818] env[61867]: INFO nova.compute.claims [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 627.805963] env[61867]: DEBUG nova.compute.manager [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Build of instance 736f861f-d9a5-4fa5-a28d-26604ebbf051 was re-scheduled: Binding failed for port 06047bd3-a41f-45f1-9e3f-7b129fd2a39a, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 627.806412] env[61867]: DEBUG nova.compute.manager [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 627.806636] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Acquiring lock "refresh_cache-736f861f-d9a5-4fa5-a28d-26604ebbf051" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.806784] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Acquired lock "refresh_cache-736f861f-d9a5-4fa5-a28d-26604ebbf051" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.806943] env[61867]: DEBUG nova.network.neutron [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 627.865616] env[61867]: DEBUG nova.compute.manager [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 627.886197] env[61867]: DEBUG nova.virt.hardware [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 627.886458] env[61867]: DEBUG nova.virt.hardware [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 627.886618] env[61867]: DEBUG nova.virt.hardware [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 627.886800] env[61867]: DEBUG nova.virt.hardware [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 627.887107] env[61867]: DEBUG nova.virt.hardware [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 627.887295] env[61867]: DEBUG nova.virt.hardware [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 627.887515] env[61867]: DEBUG nova.virt.hardware [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 627.887675] env[61867]: DEBUG nova.virt.hardware [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 627.887883] env[61867]: DEBUG nova.virt.hardware [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 627.888059] env[61867]: DEBUG nova.virt.hardware [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 627.888235] env[61867]: DEBUG nova.virt.hardware [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 627.889154] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7196e89d-ac6d-4f23-8c8c-aae1c60ce103 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.897326] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7418b8f9-3437-42ca-a6ea-41046047ab81 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.911347] env[61867]: ERROR nova.compute.manager [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4f1af8bd-6c9c-433f-8671-5beba8b94491, please check neutron logs for more information. [ 627.911347] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Traceback (most recent call last): [ 627.911347] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 627.911347] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] yield resources [ 627.911347] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 627.911347] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] self.driver.spawn(context, instance, image_meta, [ 627.911347] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 627.911347] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.911347] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.911347] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] vm_ref = self.build_virtual_machine(instance, [ 627.911347] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.911651] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.911651] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.911651] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] for vif in network_info: [ 627.911651] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.911651] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] return self._sync_wrapper(fn, *args, **kwargs) [ 627.911651] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.911651] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] self.wait() [ 627.911651] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.911651] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] self[:] = self._gt.wait() [ 627.911651] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.911651] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] return self._exit_event.wait() [ 627.911651] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 627.911651] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] current.throw(*self._exc) [ 627.911951] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.911951] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] result = function(*args, **kwargs) [ 627.911951] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.911951] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] return func(*args, **kwargs) [ 627.911951] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.911951] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] raise e [ 627.911951] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.911951] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] nwinfo = self.network_api.allocate_for_instance( [ 627.911951] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.911951] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] created_port_ids = self._update_ports_for_instance( [ 627.911951] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.911951] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] with excutils.save_and_reraise_exception(): [ 627.911951] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.912283] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] self.force_reraise() [ 627.912283] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.912283] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] raise self.value [ 627.912283] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.912283] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] updated_port = self._update_port( [ 627.912283] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.912283] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] _ensure_no_port_binding_failure(port) [ 627.912283] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.912283] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] raise exception.PortBindingFailed(port_id=port['id']) [ 627.912283] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] nova.exception.PortBindingFailed: Binding failed for port 4f1af8bd-6c9c-433f-8671-5beba8b94491, please check neutron logs for more information. [ 627.912283] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] [ 627.912283] env[61867]: INFO nova.compute.manager [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Terminating instance [ 627.913743] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Acquiring lock "refresh_cache-9ffe4fa1-4239-435b-ac7e-648cec92af65" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.968184] env[61867]: DEBUG nova.network.neutron [req-37aa9944-2f3c-4609-8e84-e21e6bd42e9f req-ad2a82b5-c421-41b4-a67a-5c29ffe4a464 service nova] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.058187] env[61867]: DEBUG nova.network.neutron [req-37aa9944-2f3c-4609-8e84-e21e6bd42e9f req-ad2a82b5-c421-41b4-a67a-5c29ffe4a464 service nova] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.064033] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.326089] env[61867]: DEBUG nova.network.neutron [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.396333] env[61867]: DEBUG nova.network.neutron [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.561630] env[61867]: DEBUG oslo_concurrency.lockutils [req-37aa9944-2f3c-4609-8e84-e21e6bd42e9f req-ad2a82b5-c421-41b4-a67a-5c29ffe4a464 service nova] Releasing lock "refresh_cache-9ffe4fa1-4239-435b-ac7e-648cec92af65" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.563018] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Acquired lock "refresh_cache-9ffe4fa1-4239-435b-ac7e-648cec92af65" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.563018] env[61867]: DEBUG nova.network.neutron [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 628.637733] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "b5a6ff7c-2b48-4f82-ba53-ec1977736c8f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.638114] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "b5a6ff7c-2b48-4f82-ba53-ec1977736c8f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.671724] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "81ada1e9-b100-44fd-8119-7ef8695e601f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.671906] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "81ada1e9-b100-44fd-8119-7ef8695e601f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.902326] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Releasing lock "refresh_cache-736f861f-d9a5-4fa5-a28d-26604ebbf051" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.902568] env[61867]: DEBUG nova.compute.manager [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 628.902749] env[61867]: DEBUG nova.compute.manager [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 628.902915] env[61867]: DEBUG nova.network.neutron [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 628.920824] env[61867]: DEBUG nova.network.neutron [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.080174] env[61867]: DEBUG nova.network.neutron [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.168017] env[61867]: DEBUG nova.network.neutron [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.174928] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b7c95d-2f9b-4108-bc45-c4c77a149078 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.183332] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd42a2a4-e869-4060-aac8-693d05914edf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.212670] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5cc0cba-924f-4dc2-9b64-64fd346eb605 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.220519] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf2f594-4bfc-4d0f-a1e7-191ff14c48fa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.233577] env[61867]: DEBUG nova.compute.provider_tree [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.423998] env[61867]: DEBUG nova.network.neutron [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.508390] env[61867]: DEBUG nova.compute.manager [req-67bff7ae-7b92-415d-9fc0-87d10efcc869 req-134cb6a3-3c30-496a-a1f4-d113eb97b1fb service nova] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Received event network-vif-deleted-4f1af8bd-6c9c-433f-8671-5beba8b94491 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 629.670782] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Releasing lock "refresh_cache-9ffe4fa1-4239-435b-ac7e-648cec92af65" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.671487] env[61867]: DEBUG nova.compute.manager [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 629.671805] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 629.672307] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-feb4e346-7cf6-4f66-a7de-0f11a7e5dbfb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.684642] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc899cfb-25b5-4ec2-ae19-599cf9f2ee80 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.710239] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9ffe4fa1-4239-435b-ac7e-648cec92af65 could not be found. [ 629.710511] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 629.710618] env[61867]: INFO nova.compute.manager [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Took 0.04 seconds to destroy the instance on the hypervisor. [ 629.710860] env[61867]: DEBUG oslo.service.loopingcall [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 629.711083] env[61867]: DEBUG nova.compute.manager [-] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 629.711175] env[61867]: DEBUG nova.network.neutron [-] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 629.725441] env[61867]: DEBUG nova.network.neutron [-] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.736696] env[61867]: DEBUG nova.scheduler.client.report [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.927218] env[61867]: INFO nova.compute.manager [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] [instance: 736f861f-d9a5-4fa5-a28d-26604ebbf051] Took 1.02 seconds to deallocate network for instance. [ 630.229054] env[61867]: DEBUG nova.network.neutron [-] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.241882] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.241882] env[61867]: DEBUG nova.compute.manager [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 630.246973] env[61867]: DEBUG oslo_concurrency.lockutils [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.046s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.731084] env[61867]: INFO nova.compute.manager [-] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Took 1.02 seconds to deallocate network for instance. [ 630.733444] env[61867]: DEBUG nova.compute.claims [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 630.733623] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.749049] env[61867]: DEBUG nova.compute.utils [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 630.752544] env[61867]: DEBUG nova.compute.manager [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 630.752780] env[61867]: DEBUG nova.network.neutron [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 630.792809] env[61867]: DEBUG nova.policy [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae1ff5effba94a5883456b6115e6c5d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f840df7fdbbd499cb6a10bf63ad3c837', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 630.958533] env[61867]: INFO nova.scheduler.client.report [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Deleted allocations for instance 736f861f-d9a5-4fa5-a28d-26604ebbf051 [ 631.063284] env[61867]: DEBUG nova.network.neutron [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Successfully created port: 58eb12aa-0864-41c7-88da-3f6eb2716588 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 631.168950] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e2bee8-2df9-4e8c-88c8-0adec830092f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.176412] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc44ac4-ffdc-4fd6-8cd5-622c617412bf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.206721] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25cbcc26-6784-441b-b204-bc6fc57dabd8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.214679] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-355e2dac-d501-4f79-950b-f41f562ac1a7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.227633] env[61867]: DEBUG nova.compute.provider_tree [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.256420] env[61867]: DEBUG nova.compute.manager [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 631.467530] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ff525b4b-1ade-4cd9-9ffb-7edf44b81b24 tempest-AttachInterfacesUnderV243Test-1904231511 tempest-AttachInterfacesUnderV243Test-1904231511-project-member] Lock "736f861f-d9a5-4fa5-a28d-26604ebbf051" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.338s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.732898] env[61867]: DEBUG nova.scheduler.client.report [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 631.741604] env[61867]: DEBUG nova.compute.manager [req-aa284aeb-b7ac-44fe-bbba-64a98ba10d44 req-73d3229d-0945-4019-bbdf-10054456f7ea service nova] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Received event network-changed-58eb12aa-0864-41c7-88da-3f6eb2716588 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 631.741726] env[61867]: DEBUG nova.compute.manager [req-aa284aeb-b7ac-44fe-bbba-64a98ba10d44 req-73d3229d-0945-4019-bbdf-10054456f7ea service nova] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Refreshing instance network info cache due to event network-changed-58eb12aa-0864-41c7-88da-3f6eb2716588. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 631.742277] env[61867]: DEBUG oslo_concurrency.lockutils [req-aa284aeb-b7ac-44fe-bbba-64a98ba10d44 req-73d3229d-0945-4019-bbdf-10054456f7ea service nova] Acquiring lock "refresh_cache-c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.742277] env[61867]: DEBUG oslo_concurrency.lockutils [req-aa284aeb-b7ac-44fe-bbba-64a98ba10d44 req-73d3229d-0945-4019-bbdf-10054456f7ea service nova] Acquired lock "refresh_cache-c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.742277] env[61867]: DEBUG nova.network.neutron [req-aa284aeb-b7ac-44fe-bbba-64a98ba10d44 req-73d3229d-0945-4019-bbdf-10054456f7ea service nova] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Refreshing network info cache for port 58eb12aa-0864-41c7-88da-3f6eb2716588 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 631.877940] env[61867]: ERROR nova.compute.manager [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 58eb12aa-0864-41c7-88da-3f6eb2716588, please check neutron logs for more information. [ 631.877940] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 631.877940] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.877940] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 631.877940] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.877940] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 631.877940] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.877940] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 631.877940] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.877940] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 631.877940] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.877940] env[61867]: ERROR nova.compute.manager raise self.value [ 631.877940] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.877940] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 631.877940] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.877940] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 631.878447] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.878447] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 631.878447] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 58eb12aa-0864-41c7-88da-3f6eb2716588, please check neutron logs for more information. [ 631.878447] env[61867]: ERROR nova.compute.manager [ 631.878447] env[61867]: Traceback (most recent call last): [ 631.878447] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 631.878447] env[61867]: listener.cb(fileno) [ 631.878447] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.878447] env[61867]: result = function(*args, **kwargs) [ 631.878447] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.878447] env[61867]: return func(*args, **kwargs) [ 631.878447] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.878447] env[61867]: raise e [ 631.878447] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.878447] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 631.878447] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.878447] env[61867]: created_port_ids = self._update_ports_for_instance( [ 631.878447] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.878447] env[61867]: with excutils.save_and_reraise_exception(): [ 631.878447] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.878447] env[61867]: self.force_reraise() [ 631.878447] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.878447] env[61867]: raise self.value [ 631.878447] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.878447] env[61867]: updated_port = self._update_port( [ 631.878447] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.878447] env[61867]: _ensure_no_port_binding_failure(port) [ 631.878447] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.878447] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 631.879285] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 58eb12aa-0864-41c7-88da-3f6eb2716588, please check neutron logs for more information. [ 631.879285] env[61867]: Removing descriptor: 17 [ 631.971960] env[61867]: DEBUG nova.compute.manager [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 632.244293] env[61867]: DEBUG oslo_concurrency.lockutils [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.997s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.244293] env[61867]: ERROR nova.compute.manager [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 984de365-ba36-4070-85a5-f2bbc4b223f6, please check neutron logs for more information. [ 632.244293] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Traceback (most recent call last): [ 632.244293] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 632.244293] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] self.driver.spawn(context, instance, image_meta, [ 632.244293] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 632.244293] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.244293] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.244293] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] vm_ref = self.build_virtual_machine(instance, [ 632.244591] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.244591] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.244591] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.244591] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] for vif in network_info: [ 632.244591] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.244591] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] return self._sync_wrapper(fn, *args, **kwargs) [ 632.244591] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.244591] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] self.wait() [ 632.244591] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.244591] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] self[:] = self._gt.wait() [ 632.244591] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.244591] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] return self._exit_event.wait() [ 632.244591] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 632.244893] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] current.throw(*self._exc) [ 632.244893] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.244893] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] result = function(*args, **kwargs) [ 632.244893] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.244893] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] return func(*args, **kwargs) [ 632.244893] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.244893] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] raise e [ 632.244893] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.244893] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] nwinfo = self.network_api.allocate_for_instance( [ 632.244893] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.244893] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] created_port_ids = self._update_ports_for_instance( [ 632.244893] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.244893] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] with excutils.save_and_reraise_exception(): [ 632.245283] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.245283] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] self.force_reraise() [ 632.245283] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.245283] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] raise self.value [ 632.245283] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.245283] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] updated_port = self._update_port( [ 632.245283] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.245283] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] _ensure_no_port_binding_failure(port) [ 632.245283] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.245283] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] raise exception.PortBindingFailed(port_id=port['id']) [ 632.245283] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] nova.exception.PortBindingFailed: Binding failed for port 984de365-ba36-4070-85a5-f2bbc4b223f6, please check neutron logs for more information. [ 632.245283] env[61867]: ERROR nova.compute.manager [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] [ 632.245621] env[61867]: DEBUG nova.compute.utils [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Binding failed for port 984de365-ba36-4070-85a5-f2bbc4b223f6, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 632.245621] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.314s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.251529] env[61867]: DEBUG nova.compute.manager [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Build of instance d25df58d-4b66-4aa4-9d07-205f36e61775 was re-scheduled: Binding failed for port 984de365-ba36-4070-85a5-f2bbc4b223f6, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 632.251991] env[61867]: DEBUG nova.compute.manager [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 632.252194] env[61867]: DEBUG oslo_concurrency.lockutils [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Acquiring lock "refresh_cache-d25df58d-4b66-4aa4-9d07-205f36e61775" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.252332] env[61867]: DEBUG oslo_concurrency.lockutils [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Acquired lock "refresh_cache-d25df58d-4b66-4aa4-9d07-205f36e61775" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.253218] env[61867]: DEBUG nova.network.neutron [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 632.265464] env[61867]: DEBUG nova.compute.manager [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 632.273487] env[61867]: DEBUG nova.network.neutron [req-aa284aeb-b7ac-44fe-bbba-64a98ba10d44 req-73d3229d-0945-4019-bbdf-10054456f7ea service nova] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.310238] env[61867]: DEBUG nova.virt.hardware [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 632.310495] env[61867]: DEBUG nova.virt.hardware [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 632.310643] env[61867]: DEBUG nova.virt.hardware [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 632.310818] env[61867]: DEBUG nova.virt.hardware [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 632.311020] env[61867]: DEBUG nova.virt.hardware [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 632.311116] env[61867]: DEBUG nova.virt.hardware [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 632.311322] env[61867]: DEBUG nova.virt.hardware [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 632.311585] env[61867]: DEBUG nova.virt.hardware [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 632.311648] env[61867]: DEBUG nova.virt.hardware [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 632.311784] env[61867]: DEBUG nova.virt.hardware [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 632.311953] env[61867]: DEBUG nova.virt.hardware [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 632.313107] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04f8f59-27c8-49e9-9c87-fbf93abfb66b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.321082] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9049379-7316-4e73-bf4a-67eb63db7574 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.335441] env[61867]: ERROR nova.compute.manager [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 58eb12aa-0864-41c7-88da-3f6eb2716588, please check neutron logs for more information. [ 632.335441] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Traceback (most recent call last): [ 632.335441] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 632.335441] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] yield resources [ 632.335441] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 632.335441] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] self.driver.spawn(context, instance, image_meta, [ 632.335441] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 632.335441] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.335441] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.335441] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] vm_ref = self.build_virtual_machine(instance, [ 632.335441] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.335812] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.335812] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.335812] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] for vif in network_info: [ 632.335812] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.335812] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] return self._sync_wrapper(fn, *args, **kwargs) [ 632.335812] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.335812] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] self.wait() [ 632.335812] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.335812] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] self[:] = self._gt.wait() [ 632.335812] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.335812] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] return self._exit_event.wait() [ 632.335812] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 632.335812] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] current.throw(*self._exc) [ 632.336165] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.336165] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] result = function(*args, **kwargs) [ 632.336165] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.336165] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] return func(*args, **kwargs) [ 632.336165] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.336165] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] raise e [ 632.336165] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.336165] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] nwinfo = self.network_api.allocate_for_instance( [ 632.336165] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.336165] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] created_port_ids = self._update_ports_for_instance( [ 632.336165] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.336165] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] with excutils.save_and_reraise_exception(): [ 632.336165] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.336544] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] self.force_reraise() [ 632.336544] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.336544] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] raise self.value [ 632.336544] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.336544] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] updated_port = self._update_port( [ 632.336544] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.336544] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] _ensure_no_port_binding_failure(port) [ 632.336544] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.336544] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] raise exception.PortBindingFailed(port_id=port['id']) [ 632.336544] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] nova.exception.PortBindingFailed: Binding failed for port 58eb12aa-0864-41c7-88da-3f6eb2716588, please check neutron logs for more information. [ 632.336544] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] [ 632.336544] env[61867]: INFO nova.compute.manager [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Terminating instance [ 632.337628] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquiring lock "refresh_cache-c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.411106] env[61867]: DEBUG nova.network.neutron [req-aa284aeb-b7ac-44fe-bbba-64a98ba10d44 req-73d3229d-0945-4019-bbdf-10054456f7ea service nova] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.495266] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.773378] env[61867]: DEBUG nova.network.neutron [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.914696] env[61867]: DEBUG oslo_concurrency.lockutils [req-aa284aeb-b7ac-44fe-bbba-64a98ba10d44 req-73d3229d-0945-4019-bbdf-10054456f7ea service nova] Releasing lock "refresh_cache-c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.915176] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquired lock "refresh_cache-c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.915441] env[61867]: DEBUG nova.network.neutron [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 633.026805] env[61867]: DEBUG nova.network.neutron [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.196553] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db865c1-faed-42c0-9f71-60031cd0a7ce {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.204997] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5e2230-1c7e-4f44-bbc0-61bca4f863ac {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.239046] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-763952c2-5584-417d-8042-7bfdac9b6f9c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.246517] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34758452-349a-4b23-807e-ffe58c77ea8e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.261106] env[61867]: DEBUG nova.compute.provider_tree [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.444945] env[61867]: DEBUG nova.network.neutron [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.530626] env[61867]: DEBUG oslo_concurrency.lockutils [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Releasing lock "refresh_cache-d25df58d-4b66-4aa4-9d07-205f36e61775" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.530851] env[61867]: DEBUG nova.compute.manager [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 633.531300] env[61867]: DEBUG nova.compute.manager [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.531495] env[61867]: DEBUG nova.network.neutron [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 633.557145] env[61867]: DEBUG nova.network.neutron [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.565305] env[61867]: DEBUG nova.network.neutron [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.768098] env[61867]: DEBUG nova.scheduler.client.report [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 633.823497] env[61867]: DEBUG nova.compute.manager [req-458ae895-0879-4915-90b5-ec364a276346 req-963aa8e6-59c8-47be-b2cd-7375d8abe67f service nova] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Received event network-vif-deleted-58eb12aa-0864-41c7-88da-3f6eb2716588 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 634.060552] env[61867]: DEBUG nova.network.neutron [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.071020] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Releasing lock "refresh_cache-c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.071020] env[61867]: DEBUG nova.compute.manager [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 634.071020] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 634.071020] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-77f59c8b-ebf7-46a7-898d-e027c74be40f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.080249] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e4716a9-7cdc-453e-a514-40e77bd3d321 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.102207] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a could not be found. [ 634.102446] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 634.102637] env[61867]: INFO nova.compute.manager [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 634.102882] env[61867]: DEBUG oslo.service.loopingcall [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 634.103103] env[61867]: DEBUG nova.compute.manager [-] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.103203] env[61867]: DEBUG nova.network.neutron [-] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 634.118242] env[61867]: DEBUG nova.network.neutron [-] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.275806] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.031s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.275806] env[61867]: ERROR nova.compute.manager [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6d020051-3a6c-4f84-922b-85cfee5b7411, please check neutron logs for more information. [ 634.275806] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Traceback (most recent call last): [ 634.275806] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 634.275806] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] self.driver.spawn(context, instance, image_meta, [ 634.275806] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 634.275806] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.275806] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.275806] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] vm_ref = self.build_virtual_machine(instance, [ 634.276120] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.276120] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.276120] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.276120] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] for vif in network_info: [ 634.276120] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 634.276120] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] return self._sync_wrapper(fn, *args, **kwargs) [ 634.276120] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 634.276120] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] self.wait() [ 634.276120] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 634.276120] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] self[:] = self._gt.wait() [ 634.276120] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.276120] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] return self._exit_event.wait() [ 634.276120] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 634.276538] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] result = hub.switch() [ 634.276538] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 634.276538] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] return self.greenlet.switch() [ 634.276538] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.276538] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] result = function(*args, **kwargs) [ 634.276538] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.276538] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] return func(*args, **kwargs) [ 634.276538] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.276538] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] raise e [ 634.276538] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.276538] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] nwinfo = self.network_api.allocate_for_instance( [ 634.276538] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.276538] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] created_port_ids = self._update_ports_for_instance( [ 634.277012] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.277012] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] with excutils.save_and_reraise_exception(): [ 634.277012] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.277012] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] self.force_reraise() [ 634.277012] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.277012] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] raise self.value [ 634.277012] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.277012] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] updated_port = self._update_port( [ 634.277012] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.277012] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] _ensure_no_port_binding_failure(port) [ 634.277012] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.277012] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] raise exception.PortBindingFailed(port_id=port['id']) [ 634.277453] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] nova.exception.PortBindingFailed: Binding failed for port 6d020051-3a6c-4f84-922b-85cfee5b7411, please check neutron logs for more information. [ 634.277453] env[61867]: ERROR nova.compute.manager [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] [ 634.277453] env[61867]: DEBUG nova.compute.utils [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Binding failed for port 6d020051-3a6c-4f84-922b-85cfee5b7411, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 634.277814] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 17.290s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.279972] env[61867]: DEBUG nova.compute.manager [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Build of instance ba4cc401-7ae3-40bc-996c-13741a65d991 was re-scheduled: Binding failed for port 6d020051-3a6c-4f84-922b-85cfee5b7411, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 634.280514] env[61867]: DEBUG nova.compute.manager [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 634.280771] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Acquiring lock "refresh_cache-ba4cc401-7ae3-40bc-996c-13741a65d991" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.280951] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Acquired lock "refresh_cache-ba4cc401-7ae3-40bc-996c-13741a65d991" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.281154] env[61867]: DEBUG nova.network.neutron [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 634.564699] env[61867]: INFO nova.compute.manager [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] [instance: d25df58d-4b66-4aa4-9d07-205f36e61775] Took 1.03 seconds to deallocate network for instance. [ 634.621415] env[61867]: DEBUG nova.network.neutron [-] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.803410] env[61867]: DEBUG nova.network.neutron [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.887740] env[61867]: DEBUG nova.network.neutron [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.126638] env[61867]: INFO nova.compute.manager [-] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Took 1.02 seconds to deallocate network for instance. [ 635.129489] env[61867]: DEBUG nova.compute.claims [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 635.130072] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.390421] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Releasing lock "refresh_cache-ba4cc401-7ae3-40bc-996c-13741a65d991" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.390620] env[61867]: DEBUG nova.compute.manager [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 635.390804] env[61867]: DEBUG nova.compute.manager [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 635.390966] env[61867]: DEBUG nova.network.neutron [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 635.407165] env[61867]: DEBUG nova.network.neutron [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.593116] env[61867]: INFO nova.scheduler.client.report [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Deleted allocations for instance d25df58d-4b66-4aa4-9d07-205f36e61775 [ 635.816563] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance d25df58d-4b66-4aa4-9d07-205f36e61775 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 635.909475] env[61867]: DEBUG nova.network.neutron [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.101645] env[61867]: DEBUG oslo_concurrency.lockutils [None req-897449fa-af6c-40b0-b355-228cd513b847 tempest-ServersTestJSON-644124255 tempest-ServersTestJSON-644124255-project-member] Lock "d25df58d-4b66-4aa4-9d07-205f36e61775" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.851s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.320444] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance ba4cc401-7ae3-40bc-996c-13741a65d991 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 636.320444] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 013fc4b2-95f7-439e-b82b-da2265ccf736 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 636.320740] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 32e947ce-c7e5-4568-b1e4-89dfc6a8009e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 636.320740] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 9ffe4fa1-4239-435b-ac7e-648cec92af65 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 636.320740] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 636.414650] env[61867]: INFO nova.compute.manager [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] [instance: ba4cc401-7ae3-40bc-996c-13741a65d991] Took 1.02 seconds to deallocate network for instance. [ 636.607193] env[61867]: DEBUG nova.compute.manager [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 636.826579] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 8f8ea2ec-31fe-4726-8f84-89251c138d44 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 637.131809] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.330588] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 50930c54-d03e-4529-ac85-a73e9a5b4e4d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 637.445721] env[61867]: INFO nova.scheduler.client.report [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Deleted allocations for instance ba4cc401-7ae3-40bc-996c-13741a65d991 [ 637.836450] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 572dbc25-39a6-4e8b-b5c0-176ea0e27839 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 637.955566] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d76bd1a0-3240-417d-84b4-2d2584d15877 tempest-ServersTestManualDisk-1244743208 tempest-ServersTestManualDisk-1244743208-project-member] Lock "ba4cc401-7ae3-40bc-996c-13741a65d991" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.853s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.339665] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 274c5625-eb81-45d1-bd95-7336bddfad1e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 638.457922] env[61867]: DEBUG nova.compute.manager [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 638.843963] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance aad1e5ea-5a9c-4499-8f67-2516c5c285cc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 638.982510] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.347511] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 639.851781] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 5c250675-3708-4c25-b370-6707c457357a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 640.354391] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 640.856792] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 59d2ee3a-756f-4dda-a70a-6d43aaf32049 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 641.360183] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 774e7c38-426a-441d-8252-3daa424be558 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 641.864144] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 69d292d9-2330-47ac-94d4-6797abb8c167 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 642.366373] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 3f1696b0-4de4-4128-bc83-e539e48dc8e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 642.869049] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 643.373822] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance add3bbbb-fd30-49ad-b95e-601af5d790e4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 643.877041] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 5dab4fcf-1a5e-493a-842b-b652e99d67ba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.380716] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance ab18091f-bab5-4926-b7a0-1f0e005bebe0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.883941] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance b1ec32c5-0642-4dca-ad43-05a5172d04d7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 645.387418] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 8656c3b8-b7b1-442a-9719-1588ee83d19b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 645.890547] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 5e97f142-b5c0-4318-9a35-085234d1cb64 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 646.393907] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 796ab4d3-ea9a-495d-9807-7780b1a7b0d6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 646.898431] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.400376] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.903369] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.405900] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance d0db58a5-d773-4e45-a238-07cfc156bdd2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.908816] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance b5a6ff7c-2b48-4f82-ba53-ec1977736c8f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.411879] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 81ada1e9-b100-44fd-8119-7ef8695e601f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.412196] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 649.412372] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 649.740981] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c73f6d-8bfa-4d85-80ea-77a1b590a684 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.748558] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd4942a2-77ed-4241-abe0-f78843cc81f0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.778137] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e469555f-f357-42d5-8eb3-39341dadad5b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.784903] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3e84d56-03ee-4e3f-bfca-a863f9c27780 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.797415] env[61867]: DEBUG nova.compute.provider_tree [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.301391] env[61867]: DEBUG nova.scheduler.client.report [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 650.809067] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61867) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 650.809067] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 16.531s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.809412] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.665s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.810994] env[61867]: INFO nova.compute.claims [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 652.145644] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a401f0dd-180c-4857-9348-9b38f83fa6ae {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.153282] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11eaa705-a461-42af-a42a-26925dff143e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.182302] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb641303-001a-4556-8711-f674cd6271ef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.189689] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8471939-40b7-4c2b-a785-66b1ae765595 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.206088] env[61867]: DEBUG nova.compute.provider_tree [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.710427] env[61867]: DEBUG nova.scheduler.client.report [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 653.214972] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.405s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.215532] env[61867]: DEBUG nova.compute.manager [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 653.218305] env[61867]: DEBUG oslo_concurrency.lockutils [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.325s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.723300] env[61867]: DEBUG nova.compute.utils [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 653.727832] env[61867]: DEBUG nova.compute.manager [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 653.728143] env[61867]: DEBUG nova.network.neutron [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 653.785085] env[61867]: DEBUG nova.policy [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1b1de18d98964d76b23bf42b4d38ad81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '027eb19757004172b9634329927c49e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 654.104611] env[61867]: DEBUG nova.network.neutron [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Successfully created port: b80e834f-0219-4832-99f4-94787663a723 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 654.109384] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84915c3e-4da5-401b-8533-04d37120c20e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.116995] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad13390d-6716-4939-bfb7-fa3acd2c3982 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.147216] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1d0264-9743-4c12-8435-9dc213ecc55e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.154699] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c212e2c6-e967-4e7b-a92e-3ba4f8193717 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.169406] env[61867]: DEBUG nova.compute.provider_tree [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.231512] env[61867]: DEBUG nova.compute.manager [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 654.692412] env[61867]: DEBUG nova.scheduler.client.report [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 654.734833] env[61867]: DEBUG nova.compute.manager [req-a7acf9a7-650e-4e3a-b2db-6a2f6090ab02 req-09726595-1cf1-447d-9cc5-7d973e33e6df service nova] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Received event network-changed-b80e834f-0219-4832-99f4-94787663a723 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 654.735047] env[61867]: DEBUG nova.compute.manager [req-a7acf9a7-650e-4e3a-b2db-6a2f6090ab02 req-09726595-1cf1-447d-9cc5-7d973e33e6df service nova] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Refreshing instance network info cache due to event network-changed-b80e834f-0219-4832-99f4-94787663a723. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 654.735274] env[61867]: DEBUG oslo_concurrency.lockutils [req-a7acf9a7-650e-4e3a-b2db-6a2f6090ab02 req-09726595-1cf1-447d-9cc5-7d973e33e6df service nova] Acquiring lock "refresh_cache-8f8ea2ec-31fe-4726-8f84-89251c138d44" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.735447] env[61867]: DEBUG oslo_concurrency.lockutils [req-a7acf9a7-650e-4e3a-b2db-6a2f6090ab02 req-09726595-1cf1-447d-9cc5-7d973e33e6df service nova] Acquired lock "refresh_cache-8f8ea2ec-31fe-4726-8f84-89251c138d44" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.735615] env[61867]: DEBUG nova.network.neutron [req-a7acf9a7-650e-4e3a-b2db-6a2f6090ab02 req-09726595-1cf1-447d-9cc5-7d973e33e6df service nova] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Refreshing network info cache for port b80e834f-0219-4832-99f4-94787663a723 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 654.901901] env[61867]: ERROR nova.compute.manager [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b80e834f-0219-4832-99f4-94787663a723, please check neutron logs for more information. [ 654.901901] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 654.901901] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.901901] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 654.901901] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.901901] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 654.901901] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.901901] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 654.901901] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.901901] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 654.901901] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.901901] env[61867]: ERROR nova.compute.manager raise self.value [ 654.901901] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.901901] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 654.901901] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.901901] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 654.902431] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.902431] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 654.902431] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b80e834f-0219-4832-99f4-94787663a723, please check neutron logs for more information. [ 654.902431] env[61867]: ERROR nova.compute.manager [ 654.902431] env[61867]: Traceback (most recent call last): [ 654.902431] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 654.902431] env[61867]: listener.cb(fileno) [ 654.902431] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.902431] env[61867]: result = function(*args, **kwargs) [ 654.902431] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 654.902431] env[61867]: return func(*args, **kwargs) [ 654.902431] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.902431] env[61867]: raise e [ 654.902431] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.902431] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 654.902431] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.902431] env[61867]: created_port_ids = self._update_ports_for_instance( [ 654.902431] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.902431] env[61867]: with excutils.save_and_reraise_exception(): [ 654.902431] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.902431] env[61867]: self.force_reraise() [ 654.902431] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.902431] env[61867]: raise self.value [ 654.902431] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.902431] env[61867]: updated_port = self._update_port( [ 654.902431] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.902431] env[61867]: _ensure_no_port_binding_failure(port) [ 654.902431] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.902431] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 654.903250] env[61867]: nova.exception.PortBindingFailed: Binding failed for port b80e834f-0219-4832-99f4-94787663a723, please check neutron logs for more information. [ 654.903250] env[61867]: Removing descriptor: 17 [ 655.178773] env[61867]: DEBUG oslo_concurrency.lockutils [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.960s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.179365] env[61867]: ERROR nova.compute.manager [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cdec95b8-f286-44cd-8d19-77aeac2b74fa, please check neutron logs for more information. [ 655.179365] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Traceback (most recent call last): [ 655.179365] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 655.179365] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] self.driver.spawn(context, instance, image_meta, [ 655.179365] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 655.179365] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] self._vmops.spawn(context, instance, image_meta, injected_files, [ 655.179365] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 655.179365] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] vm_ref = self.build_virtual_machine(instance, [ 655.179365] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 655.179365] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] vif_infos = vmwarevif.get_vif_info(self._session, [ 655.179365] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 655.179708] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] for vif in network_info: [ 655.179708] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 655.179708] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] return self._sync_wrapper(fn, *args, **kwargs) [ 655.179708] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 655.179708] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] self.wait() [ 655.179708] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 655.179708] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] self[:] = self._gt.wait() [ 655.179708] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 655.179708] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] return self._exit_event.wait() [ 655.179708] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 655.179708] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] current.throw(*self._exc) [ 655.179708] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.179708] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] result = function(*args, **kwargs) [ 655.180095] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.180095] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] return func(*args, **kwargs) [ 655.180095] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.180095] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] raise e [ 655.180095] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.180095] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] nwinfo = self.network_api.allocate_for_instance( [ 655.180095] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.180095] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] created_port_ids = self._update_ports_for_instance( [ 655.180095] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.180095] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] with excutils.save_and_reraise_exception(): [ 655.180095] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.180095] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] self.force_reraise() [ 655.180095] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.180511] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] raise self.value [ 655.180511] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.180511] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] updated_port = self._update_port( [ 655.180511] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.180511] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] _ensure_no_port_binding_failure(port) [ 655.180511] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.180511] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] raise exception.PortBindingFailed(port_id=port['id']) [ 655.180511] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] nova.exception.PortBindingFailed: Binding failed for port cdec95b8-f286-44cd-8d19-77aeac2b74fa, please check neutron logs for more information. [ 655.180511] env[61867]: ERROR nova.compute.manager [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] [ 655.180511] env[61867]: DEBUG nova.compute.utils [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Binding failed for port cdec95b8-f286-44cd-8d19-77aeac2b74fa, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 655.181331] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.538s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.184093] env[61867]: DEBUG nova.compute.manager [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Build of instance 013fc4b2-95f7-439e-b82b-da2265ccf736 was re-scheduled: Binding failed for port cdec95b8-f286-44cd-8d19-77aeac2b74fa, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 655.184509] env[61867]: DEBUG nova.compute.manager [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 655.184733] env[61867]: DEBUG oslo_concurrency.lockutils [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Acquiring lock "refresh_cache-013fc4b2-95f7-439e-b82b-da2265ccf736" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.184878] env[61867]: DEBUG oslo_concurrency.lockutils [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Acquired lock "refresh_cache-013fc4b2-95f7-439e-b82b-da2265ccf736" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.185045] env[61867]: DEBUG nova.network.neutron [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 655.243751] env[61867]: DEBUG nova.compute.manager [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 655.259408] env[61867]: DEBUG nova.network.neutron [req-a7acf9a7-650e-4e3a-b2db-6a2f6090ab02 req-09726595-1cf1-447d-9cc5-7d973e33e6df service nova] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.269454] env[61867]: DEBUG nova.virt.hardware [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 655.269708] env[61867]: DEBUG nova.virt.hardware [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 655.270132] env[61867]: DEBUG nova.virt.hardware [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 655.270240] env[61867]: DEBUG nova.virt.hardware [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 655.270342] env[61867]: DEBUG nova.virt.hardware [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 655.270667] env[61867]: DEBUG nova.virt.hardware [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 655.270957] env[61867]: DEBUG nova.virt.hardware [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 655.271178] env[61867]: DEBUG nova.virt.hardware [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 655.271372] env[61867]: DEBUG nova.virt.hardware [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 655.271538] env[61867]: DEBUG nova.virt.hardware [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 655.271708] env[61867]: DEBUG nova.virt.hardware [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 655.272574] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9e5e757-33c7-4a6c-b139-96f291bd4e6e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.280780] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23191a4e-99ed-4620-8b4a-b29894cace5c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.294384] env[61867]: ERROR nova.compute.manager [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b80e834f-0219-4832-99f4-94787663a723, please check neutron logs for more information. [ 655.294384] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Traceback (most recent call last): [ 655.294384] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 655.294384] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] yield resources [ 655.294384] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 655.294384] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] self.driver.spawn(context, instance, image_meta, [ 655.294384] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 655.294384] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] self._vmops.spawn(context, instance, image_meta, injected_files, [ 655.294384] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 655.294384] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] vm_ref = self.build_virtual_machine(instance, [ 655.294384] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 655.294813] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] vif_infos = vmwarevif.get_vif_info(self._session, [ 655.294813] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 655.294813] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] for vif in network_info: [ 655.294813] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 655.294813] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] return self._sync_wrapper(fn, *args, **kwargs) [ 655.294813] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 655.294813] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] self.wait() [ 655.294813] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 655.294813] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] self[:] = self._gt.wait() [ 655.294813] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 655.294813] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] return self._exit_event.wait() [ 655.294813] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 655.294813] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] current.throw(*self._exc) [ 655.295278] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.295278] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] result = function(*args, **kwargs) [ 655.295278] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.295278] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] return func(*args, **kwargs) [ 655.295278] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.295278] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] raise e [ 655.295278] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.295278] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] nwinfo = self.network_api.allocate_for_instance( [ 655.295278] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.295278] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] created_port_ids = self._update_ports_for_instance( [ 655.295278] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.295278] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] with excutils.save_and_reraise_exception(): [ 655.295278] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.295661] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] self.force_reraise() [ 655.295661] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.295661] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] raise self.value [ 655.295661] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.295661] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] updated_port = self._update_port( [ 655.295661] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.295661] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] _ensure_no_port_binding_failure(port) [ 655.295661] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.295661] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] raise exception.PortBindingFailed(port_id=port['id']) [ 655.295661] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] nova.exception.PortBindingFailed: Binding failed for port b80e834f-0219-4832-99f4-94787663a723, please check neutron logs for more information. [ 655.295661] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] [ 655.295661] env[61867]: INFO nova.compute.manager [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Terminating instance [ 655.297050] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Acquiring lock "refresh_cache-8f8ea2ec-31fe-4726-8f84-89251c138d44" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.341677] env[61867]: DEBUG nova.network.neutron [req-a7acf9a7-650e-4e3a-b2db-6a2f6090ab02 req-09726595-1cf1-447d-9cc5-7d973e33e6df service nova] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.708395] env[61867]: DEBUG nova.network.neutron [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.784939] env[61867]: DEBUG nova.network.neutron [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.843007] env[61867]: DEBUG oslo_concurrency.lockutils [req-a7acf9a7-650e-4e3a-b2db-6a2f6090ab02 req-09726595-1cf1-447d-9cc5-7d973e33e6df service nova] Releasing lock "refresh_cache-8f8ea2ec-31fe-4726-8f84-89251c138d44" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.843414] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Acquired lock "refresh_cache-8f8ea2ec-31fe-4726-8f84-89251c138d44" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.843598] env[61867]: DEBUG nova.network.neutron [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 656.104866] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d91c111e-b0fb-47ed-83d1-1c442191c6c0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.112211] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c34a50a-838c-436e-9b67-1f3d4ec9c4ed {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.142324] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a600376e-daa0-4995-92f3-1a1863d5b5dd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.149049] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b7015e-7c22-44ab-9a87-f73dde2b71c8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.162954] env[61867]: DEBUG nova.compute.provider_tree [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.287369] env[61867]: DEBUG oslo_concurrency.lockutils [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Releasing lock "refresh_cache-013fc4b2-95f7-439e-b82b-da2265ccf736" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.287683] env[61867]: DEBUG nova.compute.manager [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 656.287810] env[61867]: DEBUG nova.compute.manager [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 656.288022] env[61867]: DEBUG nova.network.neutron [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 656.303260] env[61867]: DEBUG nova.network.neutron [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.363221] env[61867]: DEBUG nova.network.neutron [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.437189] env[61867]: DEBUG nova.network.neutron [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.665837] env[61867]: DEBUG nova.scheduler.client.report [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 656.800011] env[61867]: DEBUG nova.compute.manager [req-a87ee587-cdf2-4f1e-ac9e-ee8a824e97bb req-7c743829-3c1b-4865-897f-212c75d73b75 service nova] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Received event network-vif-deleted-b80e834f-0219-4832-99f4-94787663a723 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 656.806990] env[61867]: DEBUG nova.network.neutron [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.940503] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Releasing lock "refresh_cache-8f8ea2ec-31fe-4726-8f84-89251c138d44" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.940950] env[61867]: DEBUG nova.compute.manager [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 656.941167] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 656.941528] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b3b73c75-dfbe-42e9-a655-db12d7b89dee {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.950364] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c994dbbc-6b0f-4b10-b19f-dcc55094b0cb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.972447] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8f8ea2ec-31fe-4726-8f84-89251c138d44 could not be found. [ 656.972668] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 656.972848] env[61867]: INFO nova.compute.manager [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Took 0.03 seconds to destroy the instance on the hypervisor. [ 656.973099] env[61867]: DEBUG oslo.service.loopingcall [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 656.973314] env[61867]: DEBUG nova.compute.manager [-] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 656.973513] env[61867]: DEBUG nova.network.neutron [-] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 656.987944] env[61867]: DEBUG nova.network.neutron [-] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.171708] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.990s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.172110] env[61867]: ERROR nova.compute.manager [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a83e557f-96a5-4bde-83ce-d197aece2270, please check neutron logs for more information. [ 657.172110] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Traceback (most recent call last): [ 657.172110] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 657.172110] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] self.driver.spawn(context, instance, image_meta, [ 657.172110] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 657.172110] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.172110] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.172110] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] vm_ref = self.build_virtual_machine(instance, [ 657.172110] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.172110] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.172110] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.172461] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] for vif in network_info: [ 657.172461] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 657.172461] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] return self._sync_wrapper(fn, *args, **kwargs) [ 657.172461] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 657.172461] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] self.wait() [ 657.172461] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 657.172461] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] self[:] = self._gt.wait() [ 657.172461] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.172461] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] return self._exit_event.wait() [ 657.172461] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 657.172461] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] current.throw(*self._exc) [ 657.172461] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.172461] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] result = function(*args, **kwargs) [ 657.172822] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.172822] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] return func(*args, **kwargs) [ 657.172822] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.172822] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] raise e [ 657.172822] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.172822] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] nwinfo = self.network_api.allocate_for_instance( [ 657.172822] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.172822] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] created_port_ids = self._update_ports_for_instance( [ 657.172822] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.172822] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] with excutils.save_and_reraise_exception(): [ 657.172822] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.172822] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] self.force_reraise() [ 657.172822] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.173208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] raise self.value [ 657.173208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.173208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] updated_port = self._update_port( [ 657.173208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.173208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] _ensure_no_port_binding_failure(port) [ 657.173208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.173208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] raise exception.PortBindingFailed(port_id=port['id']) [ 657.173208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] nova.exception.PortBindingFailed: Binding failed for port a83e557f-96a5-4bde-83ce-d197aece2270, please check neutron logs for more information. [ 657.173208] env[61867]: ERROR nova.compute.manager [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] [ 657.173208] env[61867]: DEBUG nova.compute.utils [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Binding failed for port a83e557f-96a5-4bde-83ce-d197aece2270, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 657.174033] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.025s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.175628] env[61867]: INFO nova.compute.claims [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 657.178405] env[61867]: DEBUG nova.compute.manager [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Build of instance 32e947ce-c7e5-4568-b1e4-89dfc6a8009e was re-scheduled: Binding failed for port a83e557f-96a5-4bde-83ce-d197aece2270, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 657.178830] env[61867]: DEBUG nova.compute.manager [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 657.179382] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Acquiring lock "refresh_cache-32e947ce-c7e5-4568-b1e4-89dfc6a8009e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.179382] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Acquired lock "refresh_cache-32e947ce-c7e5-4568-b1e4-89dfc6a8009e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.179382] env[61867]: DEBUG nova.network.neutron [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 657.308522] env[61867]: INFO nova.compute.manager [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] [instance: 013fc4b2-95f7-439e-b82b-da2265ccf736] Took 1.02 seconds to deallocate network for instance. [ 657.490894] env[61867]: DEBUG nova.network.neutron [-] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.701124] env[61867]: DEBUG nova.network.neutron [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.749320] env[61867]: DEBUG nova.network.neutron [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.993514] env[61867]: INFO nova.compute.manager [-] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Took 1.02 seconds to deallocate network for instance. [ 657.995961] env[61867]: DEBUG nova.compute.claims [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 657.996183] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.252109] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Releasing lock "refresh_cache-32e947ce-c7e5-4568-b1e4-89dfc6a8009e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.252348] env[61867]: DEBUG nova.compute.manager [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 658.252513] env[61867]: DEBUG nova.compute.manager [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.252683] env[61867]: DEBUG nova.network.neutron [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 658.269847] env[61867]: DEBUG nova.network.neutron [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.336458] env[61867]: INFO nova.scheduler.client.report [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Deleted allocations for instance 013fc4b2-95f7-439e-b82b-da2265ccf736 [ 658.509062] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20f85343-5224-4a8e-8986-5e6ac9504537 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.516500] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e908919c-1a44-47e8-9785-77aa45f89e13 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.546760] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-763a4cb4-82ac-44dd-b723-0fc8af643d58 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.554250] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce2e5b4-5086-4dc0-84c4-68041ce26a88 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.566690] env[61867]: DEBUG nova.compute.provider_tree [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.773057] env[61867]: DEBUG nova.network.neutron [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.845140] env[61867]: DEBUG oslo_concurrency.lockutils [None req-edc41313-a8db-47a2-8888-ef376ed5bee2 tempest-ServerActionsTestOtherA-857319036 tempest-ServerActionsTestOtherA-857319036-project-member] Lock "013fc4b2-95f7-439e-b82b-da2265ccf736" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.670s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.070625] env[61867]: DEBUG nova.scheduler.client.report [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.275202] env[61867]: INFO nova.compute.manager [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] [instance: 32e947ce-c7e5-4568-b1e4-89dfc6a8009e] Took 1.02 seconds to deallocate network for instance. [ 659.348894] env[61867]: DEBUG nova.compute.manager [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 659.576101] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.576767] env[61867]: DEBUG nova.compute.manager [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 659.582522] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.516s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.582522] env[61867]: INFO nova.compute.claims [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 659.869267] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.085982] env[61867]: DEBUG nova.compute.utils [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 660.089319] env[61867]: DEBUG nova.compute.manager [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 660.089582] env[61867]: DEBUG nova.network.neutron [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 660.143276] env[61867]: DEBUG nova.policy [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ab19047171f04596bd9515af47127a10', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '40fe7002e9fb4cc28309a019f4775c2b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 660.312392] env[61867]: INFO nova.scheduler.client.report [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Deleted allocations for instance 32e947ce-c7e5-4568-b1e4-89dfc6a8009e [ 660.585247] env[61867]: DEBUG nova.network.neutron [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Successfully created port: 91bb10e6-7994-433e-a261-e3d0ae9c2d72 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 660.592924] env[61867]: DEBUG nova.compute.manager [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 660.823922] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c42675d2-9ea9-4cd5-bd5a-8ade01d1529f tempest-ServerExternalEventsTest-865921909 tempest-ServerExternalEventsTest-865921909-project-member] Lock "32e947ce-c7e5-4568-b1e4-89dfc6a8009e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.285s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.016333] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ecd0f5f-6e87-4a59-a91c-e920975dcb52 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.026368] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff57ad5-4654-4662-a795-ccacd59ad6a0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.057907] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f954ab3-1411-4c7e-9fc5-3916654a41b8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.065348] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dfce28c-2a3c-4619-9108-327eec18631d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.079548] env[61867]: DEBUG nova.compute.provider_tree [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.329419] env[61867]: DEBUG nova.compute.manager [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 661.514243] env[61867]: ERROR nova.compute.manager [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 91bb10e6-7994-433e-a261-e3d0ae9c2d72, please check neutron logs for more information. [ 661.514243] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 661.514243] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.514243] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 661.514243] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 661.514243] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 661.514243] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 661.514243] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 661.514243] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.514243] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 661.514243] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.514243] env[61867]: ERROR nova.compute.manager raise self.value [ 661.514243] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 661.514243] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 661.514243] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.514243] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 661.514660] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.514660] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 661.514660] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 91bb10e6-7994-433e-a261-e3d0ae9c2d72, please check neutron logs for more information. [ 661.514660] env[61867]: ERROR nova.compute.manager [ 661.514660] env[61867]: Traceback (most recent call last): [ 661.514660] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 661.514660] env[61867]: listener.cb(fileno) [ 661.514660] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.514660] env[61867]: result = function(*args, **kwargs) [ 661.514660] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.514660] env[61867]: return func(*args, **kwargs) [ 661.514660] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.514660] env[61867]: raise e [ 661.514660] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.514660] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 661.514660] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 661.514660] env[61867]: created_port_ids = self._update_ports_for_instance( [ 661.514660] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 661.514660] env[61867]: with excutils.save_and_reraise_exception(): [ 661.514660] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.514660] env[61867]: self.force_reraise() [ 661.514660] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.514660] env[61867]: raise self.value [ 661.514660] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 661.514660] env[61867]: updated_port = self._update_port( [ 661.514660] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.514660] env[61867]: _ensure_no_port_binding_failure(port) [ 661.514660] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.514660] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 661.515430] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 91bb10e6-7994-433e-a261-e3d0ae9c2d72, please check neutron logs for more information. [ 661.515430] env[61867]: Removing descriptor: 17 [ 661.582745] env[61867]: DEBUG nova.scheduler.client.report [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.587681] env[61867]: DEBUG nova.compute.manager [req-b901ba53-a24f-462e-8e81-798d8d0c070f req-77850488-7ea5-4d5c-8290-899d55ca7efc service nova] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Received event network-changed-91bb10e6-7994-433e-a261-e3d0ae9c2d72 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 661.587939] env[61867]: DEBUG nova.compute.manager [req-b901ba53-a24f-462e-8e81-798d8d0c070f req-77850488-7ea5-4d5c-8290-899d55ca7efc service nova] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Refreshing instance network info cache due to event network-changed-91bb10e6-7994-433e-a261-e3d0ae9c2d72. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 661.588168] env[61867]: DEBUG oslo_concurrency.lockutils [req-b901ba53-a24f-462e-8e81-798d8d0c070f req-77850488-7ea5-4d5c-8290-899d55ca7efc service nova] Acquiring lock "refresh_cache-50930c54-d03e-4529-ac85-a73e9a5b4e4d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.588440] env[61867]: DEBUG oslo_concurrency.lockutils [req-b901ba53-a24f-462e-8e81-798d8d0c070f req-77850488-7ea5-4d5c-8290-899d55ca7efc service nova] Acquired lock "refresh_cache-50930c54-d03e-4529-ac85-a73e9a5b4e4d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.588640] env[61867]: DEBUG nova.network.neutron [req-b901ba53-a24f-462e-8e81-798d8d0c070f req-77850488-7ea5-4d5c-8290-899d55ca7efc service nova] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Refreshing network info cache for port 91bb10e6-7994-433e-a261-e3d0ae9c2d72 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 661.606795] env[61867]: DEBUG nova.compute.manager [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 661.631561] env[61867]: DEBUG nova.virt.hardware [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 661.631806] env[61867]: DEBUG nova.virt.hardware [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 661.631957] env[61867]: DEBUG nova.virt.hardware [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 661.632386] env[61867]: DEBUG nova.virt.hardware [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 661.632549] env[61867]: DEBUG nova.virt.hardware [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 661.632706] env[61867]: DEBUG nova.virt.hardware [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 661.632912] env[61867]: DEBUG nova.virt.hardware [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 661.633054] env[61867]: DEBUG nova.virt.hardware [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 661.633223] env[61867]: DEBUG nova.virt.hardware [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 661.633382] env[61867]: DEBUG nova.virt.hardware [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 661.633549] env[61867]: DEBUG nova.virt.hardware [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 661.634607] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe4ac4e2-758d-43aa-9188-e5180c45d436 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.642834] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e965482-39bc-409e-be13-2e68023259be {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.657298] env[61867]: ERROR nova.compute.manager [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 91bb10e6-7994-433e-a261-e3d0ae9c2d72, please check neutron logs for more information. [ 661.657298] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Traceback (most recent call last): [ 661.657298] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 661.657298] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] yield resources [ 661.657298] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 661.657298] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] self.driver.spawn(context, instance, image_meta, [ 661.657298] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 661.657298] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.657298] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.657298] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] vm_ref = self.build_virtual_machine(instance, [ 661.657298] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.657633] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.657633] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.657633] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] for vif in network_info: [ 661.657633] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 661.657633] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] return self._sync_wrapper(fn, *args, **kwargs) [ 661.657633] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 661.657633] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] self.wait() [ 661.657633] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 661.657633] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] self[:] = self._gt.wait() [ 661.657633] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.657633] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] return self._exit_event.wait() [ 661.657633] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 661.657633] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] current.throw(*self._exc) [ 661.657974] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.657974] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] result = function(*args, **kwargs) [ 661.657974] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.657974] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] return func(*args, **kwargs) [ 661.657974] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.657974] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] raise e [ 661.657974] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.657974] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] nwinfo = self.network_api.allocate_for_instance( [ 661.657974] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 661.657974] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] created_port_ids = self._update_ports_for_instance( [ 661.657974] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 661.657974] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] with excutils.save_and_reraise_exception(): [ 661.657974] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.658344] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] self.force_reraise() [ 661.658344] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.658344] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] raise self.value [ 661.658344] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 661.658344] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] updated_port = self._update_port( [ 661.658344] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.658344] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] _ensure_no_port_binding_failure(port) [ 661.658344] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.658344] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] raise exception.PortBindingFailed(port_id=port['id']) [ 661.658344] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] nova.exception.PortBindingFailed: Binding failed for port 91bb10e6-7994-433e-a261-e3d0ae9c2d72, please check neutron logs for more information. [ 661.658344] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] [ 661.658344] env[61867]: INFO nova.compute.manager [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Terminating instance [ 661.661758] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Acquiring lock "refresh_cache-50930c54-d03e-4529-ac85-a73e9a5b4e4d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.856680] env[61867]: DEBUG oslo_concurrency.lockutils [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.091331] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.094516] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.361s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.116609] env[61867]: DEBUG nova.network.neutron [req-b901ba53-a24f-462e-8e81-798d8d0c070f req-77850488-7ea5-4d5c-8290-899d55ca7efc service nova] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.236876] env[61867]: DEBUG nova.network.neutron [req-b901ba53-a24f-462e-8e81-798d8d0c070f req-77850488-7ea5-4d5c-8290-899d55ca7efc service nova] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.595677] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Acquiring lock "464238b5-7cd9-4212-a72a-a3ad0cd0ff15" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.595951] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Lock "464238b5-7cd9-4212-a72a-a3ad0cd0ff15" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.739792] env[61867]: DEBUG oslo_concurrency.lockutils [req-b901ba53-a24f-462e-8e81-798d8d0c070f req-77850488-7ea5-4d5c-8290-899d55ca7efc service nova] Releasing lock "refresh_cache-50930c54-d03e-4529-ac85-a73e9a5b4e4d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.744743] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Acquired lock "refresh_cache-50930c54-d03e-4529-ac85-a73e9a5b4e4d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.744743] env[61867]: DEBUG nova.network.neutron [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 662.980373] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9b26b7-6e34-4532-8b65-4b5dba98ab59 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.988617] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72626079-039f-4004-ae64-58eece148d6f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.018861] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5023c61a-c7f3-46d8-bd2d-b1a796bb851d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.026351] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d639bcfc-44d8-4654-a2d0-f565fbd5cb2d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.039414] env[61867]: DEBUG nova.compute.provider_tree [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.098711] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Lock "464238b5-7cd9-4212-a72a-a3ad0cd0ff15" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.503s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.099295] env[61867]: DEBUG nova.compute.manager [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 663.262776] env[61867]: DEBUG nova.network.neutron [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.343853] env[61867]: DEBUG nova.network.neutron [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.542583] env[61867]: DEBUG nova.scheduler.client.report [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 663.605075] env[61867]: DEBUG nova.compute.utils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 663.605450] env[61867]: DEBUG nova.compute.manager [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 663.605676] env[61867]: DEBUG nova.network.neutron [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 663.651994] env[61867]: DEBUG nova.compute.manager [req-6a3b2762-7c71-44f3-9172-28d7129d24b6 req-fd22e87c-5a78-4ee8-9410-d8c2d6d38b01 service nova] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Received event network-vif-deleted-91bb10e6-7994-433e-a261-e3d0ae9c2d72 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 663.784873] env[61867]: DEBUG nova.policy [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6c4b6a4ef7724be9a515cb18174256af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '27bb193971ae49e283d6136617a243d4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 663.850354] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Releasing lock "refresh_cache-50930c54-d03e-4529-ac85-a73e9a5b4e4d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.850902] env[61867]: DEBUG nova.compute.manager [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 663.851169] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 663.851518] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b4691976-62f8-402b-9ee4-e86699cb3a89 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.860424] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b339834-af3d-4f37-93e1-118d68c19ca3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.882175] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 50930c54-d03e-4529-ac85-a73e9a5b4e4d could not be found. [ 663.882490] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 663.882630] env[61867]: INFO nova.compute.manager [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 663.882868] env[61867]: DEBUG oslo.service.loopingcall [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 663.883094] env[61867]: DEBUG nova.compute.manager [-] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.883187] env[61867]: DEBUG nova.network.neutron [-] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 663.906029] env[61867]: DEBUG nova.network.neutron [-] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.048879] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.954s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.049535] env[61867]: ERROR nova.compute.manager [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4f1af8bd-6c9c-433f-8671-5beba8b94491, please check neutron logs for more information. [ 664.049535] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Traceback (most recent call last): [ 664.049535] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 664.049535] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] self.driver.spawn(context, instance, image_meta, [ 664.049535] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 664.049535] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] self._vmops.spawn(context, instance, image_meta, injected_files, [ 664.049535] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 664.049535] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] vm_ref = self.build_virtual_machine(instance, [ 664.049535] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 664.049535] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] vif_infos = vmwarevif.get_vif_info(self._session, [ 664.049535] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 664.049879] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] for vif in network_info: [ 664.049879] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 664.049879] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] return self._sync_wrapper(fn, *args, **kwargs) [ 664.049879] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 664.049879] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] self.wait() [ 664.049879] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 664.049879] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] self[:] = self._gt.wait() [ 664.049879] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 664.049879] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] return self._exit_event.wait() [ 664.049879] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 664.049879] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] current.throw(*self._exc) [ 664.049879] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.049879] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] result = function(*args, **kwargs) [ 664.050313] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 664.050313] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] return func(*args, **kwargs) [ 664.050313] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.050313] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] raise e [ 664.050313] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.050313] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] nwinfo = self.network_api.allocate_for_instance( [ 664.050313] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.050313] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] created_port_ids = self._update_ports_for_instance( [ 664.050313] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.050313] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] with excutils.save_and_reraise_exception(): [ 664.050313] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.050313] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] self.force_reraise() [ 664.050313] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.050756] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] raise self.value [ 664.050756] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.050756] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] updated_port = self._update_port( [ 664.050756] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.050756] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] _ensure_no_port_binding_failure(port) [ 664.050756] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.050756] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] raise exception.PortBindingFailed(port_id=port['id']) [ 664.050756] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] nova.exception.PortBindingFailed: Binding failed for port 4f1af8bd-6c9c-433f-8671-5beba8b94491, please check neutron logs for more information. [ 664.050756] env[61867]: ERROR nova.compute.manager [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] [ 664.050756] env[61867]: DEBUG nova.compute.utils [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Binding failed for port 4f1af8bd-6c9c-433f-8671-5beba8b94491, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 664.051645] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.557s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.053149] env[61867]: INFO nova.compute.claims [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 664.056432] env[61867]: DEBUG nova.compute.manager [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Build of instance 9ffe4fa1-4239-435b-ac7e-648cec92af65 was re-scheduled: Binding failed for port 4f1af8bd-6c9c-433f-8671-5beba8b94491, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 664.057088] env[61867]: DEBUG nova.compute.manager [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 664.057473] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Acquiring lock "refresh_cache-9ffe4fa1-4239-435b-ac7e-648cec92af65" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.057641] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Acquired lock "refresh_cache-9ffe4fa1-4239-435b-ac7e-648cec92af65" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.057806] env[61867]: DEBUG nova.network.neutron [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 664.060884] env[61867]: DEBUG nova.network.neutron [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Successfully created port: 783c055f-622e-4107-b61b-e923222aa69c {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 664.108760] env[61867]: DEBUG nova.compute.manager [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 664.234568] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Acquiring lock "c7203cea-2c2d-4a97-8ff4-3d33cd8443f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.234844] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Lock "c7203cea-2c2d-4a97-8ff4-3d33cd8443f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.408133] env[61867]: DEBUG nova.network.neutron [-] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.582110] env[61867]: DEBUG nova.network.neutron [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.696183] env[61867]: DEBUG nova.network.neutron [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.912357] env[61867]: INFO nova.compute.manager [-] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Took 1.03 seconds to deallocate network for instance. [ 664.915558] env[61867]: DEBUG nova.compute.claims [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 664.915744] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.963645] env[61867]: ERROR nova.compute.manager [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 783c055f-622e-4107-b61b-e923222aa69c, please check neutron logs for more information. [ 664.963645] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 664.963645] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.963645] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 664.963645] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.963645] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 664.963645] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.963645] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 664.963645] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.963645] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 664.963645] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.963645] env[61867]: ERROR nova.compute.manager raise self.value [ 664.963645] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.963645] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 664.963645] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.963645] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 664.964098] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.964098] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 664.964098] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 783c055f-622e-4107-b61b-e923222aa69c, please check neutron logs for more information. [ 664.964098] env[61867]: ERROR nova.compute.manager [ 664.964098] env[61867]: Traceback (most recent call last): [ 664.964098] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 664.964098] env[61867]: listener.cb(fileno) [ 664.964098] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.964098] env[61867]: result = function(*args, **kwargs) [ 664.964098] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 664.964098] env[61867]: return func(*args, **kwargs) [ 664.964098] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.964098] env[61867]: raise e [ 664.964098] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.964098] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 664.964098] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.964098] env[61867]: created_port_ids = self._update_ports_for_instance( [ 664.964098] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.964098] env[61867]: with excutils.save_and_reraise_exception(): [ 664.964098] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.964098] env[61867]: self.force_reraise() [ 664.964098] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.964098] env[61867]: raise self.value [ 664.964098] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.964098] env[61867]: updated_port = self._update_port( [ 664.964098] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.964098] env[61867]: _ensure_no_port_binding_failure(port) [ 664.964098] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.964098] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 664.964811] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 783c055f-622e-4107-b61b-e923222aa69c, please check neutron logs for more information. [ 664.964811] env[61867]: Removing descriptor: 17 [ 665.118671] env[61867]: DEBUG nova.compute.manager [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 665.144209] env[61867]: DEBUG nova.virt.hardware [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 665.144467] env[61867]: DEBUG nova.virt.hardware [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 665.144636] env[61867]: DEBUG nova.virt.hardware [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 665.144815] env[61867]: DEBUG nova.virt.hardware [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 665.144961] env[61867]: DEBUG nova.virt.hardware [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 665.145118] env[61867]: DEBUG nova.virt.hardware [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 665.145324] env[61867]: DEBUG nova.virt.hardware [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 665.145485] env[61867]: DEBUG nova.virt.hardware [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 665.145646] env[61867]: DEBUG nova.virt.hardware [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 665.145804] env[61867]: DEBUG nova.virt.hardware [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 665.145971] env[61867]: DEBUG nova.virt.hardware [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 665.146846] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e20979-4b58-4efe-9200-ee7e1cbb7506 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.156511] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db0183f2-420a-4847-8a0f-8055d09ee1b9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.171991] env[61867]: ERROR nova.compute.manager [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 783c055f-622e-4107-b61b-e923222aa69c, please check neutron logs for more information. [ 665.171991] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Traceback (most recent call last): [ 665.171991] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 665.171991] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] yield resources [ 665.171991] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.171991] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] self.driver.spawn(context, instance, image_meta, [ 665.171991] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 665.171991] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.171991] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.171991] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] vm_ref = self.build_virtual_machine(instance, [ 665.171991] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.172364] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.172364] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.172364] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] for vif in network_info: [ 665.172364] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.172364] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] return self._sync_wrapper(fn, *args, **kwargs) [ 665.172364] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.172364] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] self.wait() [ 665.172364] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.172364] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] self[:] = self._gt.wait() [ 665.172364] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.172364] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] return self._exit_event.wait() [ 665.172364] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 665.172364] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] current.throw(*self._exc) [ 665.172737] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.172737] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] result = function(*args, **kwargs) [ 665.172737] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.172737] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] return func(*args, **kwargs) [ 665.172737] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.172737] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] raise e [ 665.172737] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.172737] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] nwinfo = self.network_api.allocate_for_instance( [ 665.172737] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.172737] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] created_port_ids = self._update_ports_for_instance( [ 665.172737] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.172737] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] with excutils.save_and_reraise_exception(): [ 665.172737] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.173150] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] self.force_reraise() [ 665.173150] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.173150] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] raise self.value [ 665.173150] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.173150] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] updated_port = self._update_port( [ 665.173150] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.173150] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] _ensure_no_port_binding_failure(port) [ 665.173150] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.173150] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] raise exception.PortBindingFailed(port_id=port['id']) [ 665.173150] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] nova.exception.PortBindingFailed: Binding failed for port 783c055f-622e-4107-b61b-e923222aa69c, please check neutron logs for more information. [ 665.173150] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] [ 665.173150] env[61867]: INFO nova.compute.manager [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Terminating instance [ 665.175917] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Acquiring lock "refresh_cache-572dbc25-39a6-4e8b-b5c0-176ea0e27839" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.176044] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Acquired lock "refresh_cache-572dbc25-39a6-4e8b-b5c0-176ea0e27839" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.176215] env[61867]: DEBUG nova.network.neutron [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 665.199227] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Releasing lock "refresh_cache-9ffe4fa1-4239-435b-ac7e-648cec92af65" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.199441] env[61867]: DEBUG nova.compute.manager [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 665.199616] env[61867]: DEBUG nova.compute.manager [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 665.199772] env[61867]: DEBUG nova.network.neutron [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 665.217087] env[61867]: DEBUG nova.network.neutron [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.364040] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1645a1f8-d4a3-4514-acc5-59d97e8a5f32 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.371272] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6412fc2-9bd7-4334-b8a0-016650589605 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.401599] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714be599-3089-4d7b-9f79-677fd360f1fb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.408953] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e0e647-a429-4bc3-9efa-f98b0bd897b8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.422122] env[61867]: DEBUG nova.compute.provider_tree [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.693430] env[61867]: DEBUG nova.network.neutron [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.719913] env[61867]: DEBUG nova.network.neutron [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.776622] env[61867]: DEBUG nova.network.neutron [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.790752] env[61867]: DEBUG nova.compute.manager [req-16e57a9c-a3bc-47fc-9f2e-80278dfe67ac req-80a5f9c7-91b0-454d-94f1-27ffa0a81908 service nova] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Received event network-changed-783c055f-622e-4107-b61b-e923222aa69c {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 665.791009] env[61867]: DEBUG nova.compute.manager [req-16e57a9c-a3bc-47fc-9f2e-80278dfe67ac req-80a5f9c7-91b0-454d-94f1-27ffa0a81908 service nova] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Refreshing instance network info cache due to event network-changed-783c055f-622e-4107-b61b-e923222aa69c. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 665.791215] env[61867]: DEBUG oslo_concurrency.lockutils [req-16e57a9c-a3bc-47fc-9f2e-80278dfe67ac req-80a5f9c7-91b0-454d-94f1-27ffa0a81908 service nova] Acquiring lock "refresh_cache-572dbc25-39a6-4e8b-b5c0-176ea0e27839" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.925812] env[61867]: DEBUG nova.scheduler.client.report [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.223031] env[61867]: INFO nova.compute.manager [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] [instance: 9ffe4fa1-4239-435b-ac7e-648cec92af65] Took 1.02 seconds to deallocate network for instance. [ 666.279495] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Releasing lock "refresh_cache-572dbc25-39a6-4e8b-b5c0-176ea0e27839" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.280130] env[61867]: DEBUG nova.compute.manager [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 666.280352] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 666.280691] env[61867]: DEBUG oslo_concurrency.lockutils [req-16e57a9c-a3bc-47fc-9f2e-80278dfe67ac req-80a5f9c7-91b0-454d-94f1-27ffa0a81908 service nova] Acquired lock "refresh_cache-572dbc25-39a6-4e8b-b5c0-176ea0e27839" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.280871] env[61867]: DEBUG nova.network.neutron [req-16e57a9c-a3bc-47fc-9f2e-80278dfe67ac req-80a5f9c7-91b0-454d-94f1-27ffa0a81908 service nova] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Refreshing network info cache for port 783c055f-622e-4107-b61b-e923222aa69c {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 666.281959] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-04917030-818a-4bba-aa3d-a165381b272e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.291825] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b59c42b2-27f2-4748-ac78-92e4837d0411 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.314637] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 572dbc25-39a6-4e8b-b5c0-176ea0e27839 could not be found. [ 666.314860] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 666.315052] env[61867]: INFO nova.compute.manager [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Took 0.03 seconds to destroy the instance on the hypervisor. [ 666.315300] env[61867]: DEBUG oslo.service.loopingcall [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 666.315515] env[61867]: DEBUG nova.compute.manager [-] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 666.315608] env[61867]: DEBUG nova.network.neutron [-] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 666.331733] env[61867]: DEBUG nova.network.neutron [-] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.432387] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.433170] env[61867]: DEBUG nova.compute.manager [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 666.437290] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.308s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.800304] env[61867]: DEBUG nova.network.neutron [req-16e57a9c-a3bc-47fc-9f2e-80278dfe67ac req-80a5f9c7-91b0-454d-94f1-27ffa0a81908 service nova] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.834075] env[61867]: DEBUG nova.network.neutron [-] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.874947] env[61867]: DEBUG nova.network.neutron [req-16e57a9c-a3bc-47fc-9f2e-80278dfe67ac req-80a5f9c7-91b0-454d-94f1-27ffa0a81908 service nova] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.942570] env[61867]: DEBUG nova.compute.utils [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 666.946900] env[61867]: DEBUG nova.compute.manager [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 666.947099] env[61867]: DEBUG nova.network.neutron [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 666.990631] env[61867]: DEBUG nova.policy [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'adcd9eb75ecc4eccb335ebfec207a900', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7316e4f263a9432ab2f9f91484d62f58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 667.266917] env[61867]: INFO nova.scheduler.client.report [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Deleted allocations for instance 9ffe4fa1-4239-435b-ac7e-648cec92af65 [ 667.273624] env[61867]: DEBUG nova.network.neutron [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Successfully created port: 556c9e4c-9a3a-4e1f-9c99-ea0e3331ef02 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 667.337019] env[61867]: INFO nova.compute.manager [-] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Took 1.02 seconds to deallocate network for instance. [ 667.341167] env[61867]: DEBUG nova.compute.claims [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 667.341167] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.370171] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8089522-ab8b-47a9-b3e7-1e41354f96b1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.377820] env[61867]: DEBUG oslo_concurrency.lockutils [req-16e57a9c-a3bc-47fc-9f2e-80278dfe67ac req-80a5f9c7-91b0-454d-94f1-27ffa0a81908 service nova] Releasing lock "refresh_cache-572dbc25-39a6-4e8b-b5c0-176ea0e27839" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.378087] env[61867]: DEBUG nova.compute.manager [req-16e57a9c-a3bc-47fc-9f2e-80278dfe67ac req-80a5f9c7-91b0-454d-94f1-27ffa0a81908 service nova] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Received event network-vif-deleted-783c055f-622e-4107-b61b-e923222aa69c {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 667.379303] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ebeb9c-f861-4e13-9e66-07568cd443d4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.409254] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4270b76-22f9-44f1-a677-36469c64db61 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.417116] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce97657-5611-483f-bd71-86b67264209e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.435119] env[61867]: DEBUG nova.compute.provider_tree [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.450666] env[61867]: DEBUG nova.compute.manager [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 667.781223] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e679d9e-c7e3-4d07-bee2-c251f2bab8fa tempest-ImagesNegativeTestJSON-479213196 tempest-ImagesNegativeTestJSON-479213196-project-member] Lock "9ffe4fa1-4239-435b-ac7e-648cec92af65" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.023s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.937742] env[61867]: DEBUG nova.scheduler.client.report [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 667.983338] env[61867]: DEBUG nova.compute.manager [req-a033524c-9e1f-4821-b3e7-ac55b00264dc req-c7c655a0-128b-40b0-87f4-e81e4df83059 service nova] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Received event network-changed-556c9e4c-9a3a-4e1f-9c99-ea0e3331ef02 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 667.983338] env[61867]: DEBUG nova.compute.manager [req-a033524c-9e1f-4821-b3e7-ac55b00264dc req-c7c655a0-128b-40b0-87f4-e81e4df83059 service nova] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Refreshing instance network info cache due to event network-changed-556c9e4c-9a3a-4e1f-9c99-ea0e3331ef02. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 667.983338] env[61867]: DEBUG oslo_concurrency.lockutils [req-a033524c-9e1f-4821-b3e7-ac55b00264dc req-c7c655a0-128b-40b0-87f4-e81e4df83059 service nova] Acquiring lock "refresh_cache-274c5625-eb81-45d1-bd95-7336bddfad1e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.983338] env[61867]: DEBUG oslo_concurrency.lockutils [req-a033524c-9e1f-4821-b3e7-ac55b00264dc req-c7c655a0-128b-40b0-87f4-e81e4df83059 service nova] Acquired lock "refresh_cache-274c5625-eb81-45d1-bd95-7336bddfad1e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.983338] env[61867]: DEBUG nova.network.neutron [req-a033524c-9e1f-4821-b3e7-ac55b00264dc req-c7c655a0-128b-40b0-87f4-e81e4df83059 service nova] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Refreshing network info cache for port 556c9e4c-9a3a-4e1f-9c99-ea0e3331ef02 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 668.146121] env[61867]: ERROR nova.compute.manager [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 556c9e4c-9a3a-4e1f-9c99-ea0e3331ef02, please check neutron logs for more information. [ 668.146121] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 668.146121] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.146121] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 668.146121] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.146121] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 668.146121] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.146121] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 668.146121] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.146121] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 668.146121] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.146121] env[61867]: ERROR nova.compute.manager raise self.value [ 668.146121] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.146121] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 668.146121] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.146121] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 668.146599] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.146599] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 668.146599] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 556c9e4c-9a3a-4e1f-9c99-ea0e3331ef02, please check neutron logs for more information. [ 668.146599] env[61867]: ERROR nova.compute.manager [ 668.146599] env[61867]: Traceback (most recent call last): [ 668.146599] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 668.146599] env[61867]: listener.cb(fileno) [ 668.146599] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.146599] env[61867]: result = function(*args, **kwargs) [ 668.146599] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 668.146599] env[61867]: return func(*args, **kwargs) [ 668.146599] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.146599] env[61867]: raise e [ 668.146599] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.146599] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 668.146599] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.146599] env[61867]: created_port_ids = self._update_ports_for_instance( [ 668.146599] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.146599] env[61867]: with excutils.save_and_reraise_exception(): [ 668.146599] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.146599] env[61867]: self.force_reraise() [ 668.146599] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.146599] env[61867]: raise self.value [ 668.146599] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.146599] env[61867]: updated_port = self._update_port( [ 668.146599] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.146599] env[61867]: _ensure_no_port_binding_failure(port) [ 668.146599] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.146599] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 668.147501] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 556c9e4c-9a3a-4e1f-9c99-ea0e3331ef02, please check neutron logs for more information. [ 668.147501] env[61867]: Removing descriptor: 17 [ 668.283702] env[61867]: DEBUG nova.compute.manager [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 668.445019] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.005s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.445019] env[61867]: ERROR nova.compute.manager [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 58eb12aa-0864-41c7-88da-3f6eb2716588, please check neutron logs for more information. [ 668.445019] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Traceback (most recent call last): [ 668.445019] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 668.445019] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] self.driver.spawn(context, instance, image_meta, [ 668.445019] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 668.445019] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 668.445019] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 668.445019] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] vm_ref = self.build_virtual_machine(instance, [ 668.445394] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 668.445394] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] vif_infos = vmwarevif.get_vif_info(self._session, [ 668.445394] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 668.445394] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] for vif in network_info: [ 668.445394] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 668.445394] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] return self._sync_wrapper(fn, *args, **kwargs) [ 668.445394] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 668.445394] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] self.wait() [ 668.445394] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 668.445394] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] self[:] = self._gt.wait() [ 668.445394] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 668.445394] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] return self._exit_event.wait() [ 668.445394] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 668.445772] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] current.throw(*self._exc) [ 668.445772] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.445772] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] result = function(*args, **kwargs) [ 668.445772] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 668.445772] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] return func(*args, **kwargs) [ 668.445772] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.445772] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] raise e [ 668.445772] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.445772] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] nwinfo = self.network_api.allocate_for_instance( [ 668.445772] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.445772] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] created_port_ids = self._update_ports_for_instance( [ 668.445772] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.445772] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] with excutils.save_and_reraise_exception(): [ 668.446138] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.446138] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] self.force_reraise() [ 668.446138] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.446138] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] raise self.value [ 668.446138] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.446138] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] updated_port = self._update_port( [ 668.446138] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.446138] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] _ensure_no_port_binding_failure(port) [ 668.446138] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.446138] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] raise exception.PortBindingFailed(port_id=port['id']) [ 668.446138] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] nova.exception.PortBindingFailed: Binding failed for port 58eb12aa-0864-41c7-88da-3f6eb2716588, please check neutron logs for more information. [ 668.446138] env[61867]: ERROR nova.compute.manager [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] [ 668.446477] env[61867]: DEBUG nova.compute.utils [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Binding failed for port 58eb12aa-0864-41c7-88da-3f6eb2716588, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 668.446526] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.315s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.448043] env[61867]: INFO nova.compute.claims [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 668.451037] env[61867]: DEBUG nova.compute.manager [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Build of instance c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a was re-scheduled: Binding failed for port 58eb12aa-0864-41c7-88da-3f6eb2716588, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 668.451248] env[61867]: DEBUG nova.compute.manager [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 668.451473] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquiring lock "refresh_cache-c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.451617] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquired lock "refresh_cache-c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.451773] env[61867]: DEBUG nova.network.neutron [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 668.461100] env[61867]: DEBUG nova.compute.manager [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 668.487366] env[61867]: DEBUG nova.virt.hardware [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 668.487627] env[61867]: DEBUG nova.virt.hardware [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 668.487747] env[61867]: DEBUG nova.virt.hardware [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 668.487918] env[61867]: DEBUG nova.virt.hardware [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 668.488068] env[61867]: DEBUG nova.virt.hardware [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 668.488213] env[61867]: DEBUG nova.virt.hardware [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 668.488484] env[61867]: DEBUG nova.virt.hardware [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 668.488654] env[61867]: DEBUG nova.virt.hardware [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 668.488819] env[61867]: DEBUG nova.virt.hardware [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 668.488978] env[61867]: DEBUG nova.virt.hardware [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 668.489164] env[61867]: DEBUG nova.virt.hardware [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 668.491894] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c40bb4-b2dc-450f-8ef4-e895c24b3865 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.500402] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e656ba-ecc9-47bb-b696-8b0e52d056af {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.514038] env[61867]: ERROR nova.compute.manager [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 556c9e4c-9a3a-4e1f-9c99-ea0e3331ef02, please check neutron logs for more information. [ 668.514038] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Traceback (most recent call last): [ 668.514038] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 668.514038] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] yield resources [ 668.514038] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 668.514038] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] self.driver.spawn(context, instance, image_meta, [ 668.514038] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 668.514038] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 668.514038] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 668.514038] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] vm_ref = self.build_virtual_machine(instance, [ 668.514038] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 668.514372] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] vif_infos = vmwarevif.get_vif_info(self._session, [ 668.514372] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 668.514372] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] for vif in network_info: [ 668.514372] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 668.514372] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] return self._sync_wrapper(fn, *args, **kwargs) [ 668.514372] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 668.514372] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] self.wait() [ 668.514372] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 668.514372] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] self[:] = self._gt.wait() [ 668.514372] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 668.514372] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] return self._exit_event.wait() [ 668.514372] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 668.514372] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] current.throw(*self._exc) [ 668.514750] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.514750] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] result = function(*args, **kwargs) [ 668.514750] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 668.514750] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] return func(*args, **kwargs) [ 668.514750] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.514750] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] raise e [ 668.514750] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.514750] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] nwinfo = self.network_api.allocate_for_instance( [ 668.514750] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.514750] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] created_port_ids = self._update_ports_for_instance( [ 668.514750] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.514750] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] with excutils.save_and_reraise_exception(): [ 668.514750] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.515109] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] self.force_reraise() [ 668.515109] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.515109] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] raise self.value [ 668.515109] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.515109] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] updated_port = self._update_port( [ 668.515109] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.515109] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] _ensure_no_port_binding_failure(port) [ 668.515109] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.515109] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] raise exception.PortBindingFailed(port_id=port['id']) [ 668.515109] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] nova.exception.PortBindingFailed: Binding failed for port 556c9e4c-9a3a-4e1f-9c99-ea0e3331ef02, please check neutron logs for more information. [ 668.515109] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] [ 668.515109] env[61867]: INFO nova.compute.manager [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Terminating instance [ 668.516162] env[61867]: DEBUG nova.network.neutron [req-a033524c-9e1f-4821-b3e7-ac55b00264dc req-c7c655a0-128b-40b0-87f4-e81e4df83059 service nova] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.517910] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "refresh_cache-274c5625-eb81-45d1-bd95-7336bddfad1e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.628540] env[61867]: DEBUG nova.network.neutron [req-a033524c-9e1f-4821-b3e7-ac55b00264dc req-c7c655a0-128b-40b0-87f4-e81e4df83059 service nova] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.813281] env[61867]: DEBUG oslo_concurrency.lockutils [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.976467] env[61867]: DEBUG nova.network.neutron [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.108404] env[61867]: DEBUG nova.network.neutron [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.131385] env[61867]: DEBUG oslo_concurrency.lockutils [req-a033524c-9e1f-4821-b3e7-ac55b00264dc req-c7c655a0-128b-40b0-87f4-e81e4df83059 service nova] Releasing lock "refresh_cache-274c5625-eb81-45d1-bd95-7336bddfad1e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.131785] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "refresh_cache-274c5625-eb81-45d1-bd95-7336bddfad1e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.131975] env[61867]: DEBUG nova.network.neutron [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 669.617575] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Releasing lock "refresh_cache-c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.617822] env[61867]: DEBUG nova.compute.manager [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 669.618017] env[61867]: DEBUG nova.compute.manager [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 669.618189] env[61867]: DEBUG nova.network.neutron [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 669.634170] env[61867]: DEBUG nova.network.neutron [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.654035] env[61867]: DEBUG nova.network.neutron [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.744259] env[61867]: DEBUG nova.network.neutron [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.807238] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb883c6-9eda-4cee-91f3-7ef71c7c942d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.814757] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a23239-8907-41e1-a56d-a9a6b51bcadf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.847992] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451a309c-e54a-444f-aab0-7eab660e5213 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.855424] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b5df85a-1a34-4172-a88c-fdfec17734a5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.868546] env[61867]: DEBUG nova.compute.provider_tree [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.010217] env[61867]: DEBUG nova.compute.manager [req-645d900a-c10e-4f18-af6f-5febc7dacaad req-5afdc341-85ff-48d5-a840-a34f77de4df6 service nova] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Received event network-vif-deleted-556c9e4c-9a3a-4e1f-9c99-ea0e3331ef02 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 670.137881] env[61867]: DEBUG nova.network.neutron [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.246926] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "refresh_cache-274c5625-eb81-45d1-bd95-7336bddfad1e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.247648] env[61867]: DEBUG nova.compute.manager [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 670.247648] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 670.247863] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1445874f-98eb-4d80-8616-2cf0fe5486f9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.256445] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be288c48-177c-40d8-8e17-c69d97d93e93 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.278181] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 274c5625-eb81-45d1-bd95-7336bddfad1e could not be found. [ 670.278448] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 670.278636] env[61867]: INFO nova.compute.manager [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 670.278885] env[61867]: DEBUG oslo.service.loopingcall [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 670.279115] env[61867]: DEBUG nova.compute.manager [-] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 670.279208] env[61867]: DEBUG nova.network.neutron [-] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 670.295077] env[61867]: DEBUG nova.network.neutron [-] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.372343] env[61867]: DEBUG nova.scheduler.client.report [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 670.640529] env[61867]: INFO nova.compute.manager [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a] Took 1.02 seconds to deallocate network for instance. [ 670.798361] env[61867]: DEBUG nova.network.neutron [-] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.879304] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.433s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.879837] env[61867]: DEBUG nova.compute.manager [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 670.882308] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.900s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.883673] env[61867]: INFO nova.compute.claims [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 671.300549] env[61867]: INFO nova.compute.manager [-] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Took 1.02 seconds to deallocate network for instance. [ 671.302935] env[61867]: DEBUG nova.compute.claims [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 671.303220] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.387913] env[61867]: DEBUG nova.compute.utils [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 671.391172] env[61867]: DEBUG nova.compute.manager [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 671.391352] env[61867]: DEBUG nova.network.neutron [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 671.439519] env[61867]: DEBUG nova.policy [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc358009ebe6495a881034439d00978f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a29c5e9cfdaa4cc88ed300100a308ffa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 671.674711] env[61867]: INFO nova.scheduler.client.report [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Deleted allocations for instance c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a [ 671.684021] env[61867]: DEBUG nova.network.neutron [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Successfully created port: 7ab0af92-a6e0-46c0-bbf5-bda124b94276 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 671.892559] env[61867]: DEBUG nova.compute.manager [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 672.185967] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8e3208ba-e03c-4dd9-907c-b76251da2dcc tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "c50ef640-6dc2-4a8c-bbb5-cd6bc23aba6a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.168s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.346259] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a745f927-b922-4ffb-b2d5-b746ede3b297 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.360353] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-976cabf6-f7dd-41ac-a9e0-2d7e36e49090 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.396191] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb55c19a-985d-427c-a2e3-b57fbeb78fe9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.407549] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088933ec-0062-4e93-9832-c98b2d367f6b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.426743] env[61867]: DEBUG nova.compute.provider_tree [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.576688] env[61867]: ERROR nova.compute.manager [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7ab0af92-a6e0-46c0-bbf5-bda124b94276, please check neutron logs for more information. [ 672.576688] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 672.576688] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.576688] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 672.576688] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.576688] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 672.576688] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.576688] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 672.576688] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.576688] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 672.576688] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.576688] env[61867]: ERROR nova.compute.manager raise self.value [ 672.576688] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.576688] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 672.576688] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.576688] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 672.577211] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.577211] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 672.577211] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7ab0af92-a6e0-46c0-bbf5-bda124b94276, please check neutron logs for more information. [ 672.577211] env[61867]: ERROR nova.compute.manager [ 672.577211] env[61867]: Traceback (most recent call last): [ 672.577211] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 672.577211] env[61867]: listener.cb(fileno) [ 672.577211] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.577211] env[61867]: result = function(*args, **kwargs) [ 672.577211] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 672.577211] env[61867]: return func(*args, **kwargs) [ 672.577211] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.577211] env[61867]: raise e [ 672.577211] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.577211] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 672.577211] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.577211] env[61867]: created_port_ids = self._update_ports_for_instance( [ 672.577211] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.577211] env[61867]: with excutils.save_and_reraise_exception(): [ 672.577211] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.577211] env[61867]: self.force_reraise() [ 672.577211] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.577211] env[61867]: raise self.value [ 672.577211] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.577211] env[61867]: updated_port = self._update_port( [ 672.577211] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.577211] env[61867]: _ensure_no_port_binding_failure(port) [ 672.577211] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.577211] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 672.578593] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 7ab0af92-a6e0-46c0-bbf5-bda124b94276, please check neutron logs for more information. [ 672.578593] env[61867]: Removing descriptor: 17 [ 672.579387] env[61867]: DEBUG nova.compute.manager [req-fb1a607a-fc72-430c-996c-9a8c254be998 req-604fc6f3-a408-44a8-983f-1aa379f572ae service nova] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Received event network-changed-7ab0af92-a6e0-46c0-bbf5-bda124b94276 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 672.579649] env[61867]: DEBUG nova.compute.manager [req-fb1a607a-fc72-430c-996c-9a8c254be998 req-604fc6f3-a408-44a8-983f-1aa379f572ae service nova] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Refreshing instance network info cache due to event network-changed-7ab0af92-a6e0-46c0-bbf5-bda124b94276. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 672.579878] env[61867]: DEBUG oslo_concurrency.lockutils [req-fb1a607a-fc72-430c-996c-9a8c254be998 req-604fc6f3-a408-44a8-983f-1aa379f572ae service nova] Acquiring lock "refresh_cache-aad1e5ea-5a9c-4499-8f67-2516c5c285cc" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.580035] env[61867]: DEBUG oslo_concurrency.lockutils [req-fb1a607a-fc72-430c-996c-9a8c254be998 req-604fc6f3-a408-44a8-983f-1aa379f572ae service nova] Acquired lock "refresh_cache-aad1e5ea-5a9c-4499-8f67-2516c5c285cc" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.580201] env[61867]: DEBUG nova.network.neutron [req-fb1a607a-fc72-430c-996c-9a8c254be998 req-604fc6f3-a408-44a8-983f-1aa379f572ae service nova] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Refreshing network info cache for port 7ab0af92-a6e0-46c0-bbf5-bda124b94276 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 672.690573] env[61867]: DEBUG nova.compute.manager [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 672.905094] env[61867]: DEBUG nova.compute.manager [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 672.929760] env[61867]: DEBUG nova.scheduler.client.report [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.949304] env[61867]: DEBUG nova.virt.hardware [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 672.949304] env[61867]: DEBUG nova.virt.hardware [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 672.949304] env[61867]: DEBUG nova.virt.hardware [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 672.949304] env[61867]: DEBUG nova.virt.hardware [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 672.949749] env[61867]: DEBUG nova.virt.hardware [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 672.949749] env[61867]: DEBUG nova.virt.hardware [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 672.949749] env[61867]: DEBUG nova.virt.hardware [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 672.949749] env[61867]: DEBUG nova.virt.hardware [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 672.949874] env[61867]: DEBUG nova.virt.hardware [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 672.950054] env[61867]: DEBUG nova.virt.hardware [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 672.950705] env[61867]: DEBUG nova.virt.hardware [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 672.951226] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41beb72-5673-407b-acb4-e659334c835c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.959976] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5050df8c-b61e-4294-9db3-34f1bb323fea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.975632] env[61867]: ERROR nova.compute.manager [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7ab0af92-a6e0-46c0-bbf5-bda124b94276, please check neutron logs for more information. [ 672.975632] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Traceback (most recent call last): [ 672.975632] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 672.975632] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] yield resources [ 672.975632] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 672.975632] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] self.driver.spawn(context, instance, image_meta, [ 672.975632] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 672.975632] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.975632] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.975632] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] vm_ref = self.build_virtual_machine(instance, [ 672.975632] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.976167] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.976167] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.976167] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] for vif in network_info: [ 672.976167] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 672.976167] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] return self._sync_wrapper(fn, *args, **kwargs) [ 672.976167] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 672.976167] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] self.wait() [ 672.976167] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 672.976167] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] self[:] = self._gt.wait() [ 672.976167] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.976167] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] return self._exit_event.wait() [ 672.976167] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 672.976167] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] current.throw(*self._exc) [ 672.976795] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.976795] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] result = function(*args, **kwargs) [ 672.976795] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 672.976795] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] return func(*args, **kwargs) [ 672.976795] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.976795] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] raise e [ 672.976795] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.976795] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] nwinfo = self.network_api.allocate_for_instance( [ 672.976795] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.976795] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] created_port_ids = self._update_ports_for_instance( [ 672.976795] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.976795] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] with excutils.save_and_reraise_exception(): [ 672.976795] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.977178] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] self.force_reraise() [ 672.977178] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.977178] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] raise self.value [ 672.977178] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.977178] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] updated_port = self._update_port( [ 672.977178] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.977178] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] _ensure_no_port_binding_failure(port) [ 672.977178] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.977178] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] raise exception.PortBindingFailed(port_id=port['id']) [ 672.977178] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] nova.exception.PortBindingFailed: Binding failed for port 7ab0af92-a6e0-46c0-bbf5-bda124b94276, please check neutron logs for more information. [ 672.977178] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] [ 672.977178] env[61867]: INFO nova.compute.manager [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Terminating instance [ 672.978544] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "refresh_cache-aad1e5ea-5a9c-4499-8f67-2516c5c285cc" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.103481] env[61867]: DEBUG nova.network.neutron [req-fb1a607a-fc72-430c-996c-9a8c254be998 req-604fc6f3-a408-44a8-983f-1aa379f572ae service nova] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.189735] env[61867]: DEBUG nova.network.neutron [req-fb1a607a-fc72-430c-996c-9a8c254be998 req-604fc6f3-a408-44a8-983f-1aa379f572ae service nova] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.214314] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.441180] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.558s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.441416] env[61867]: DEBUG nova.compute.manager [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 673.444240] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.448s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.696703] env[61867]: DEBUG oslo_concurrency.lockutils [req-fb1a607a-fc72-430c-996c-9a8c254be998 req-604fc6f3-a408-44a8-983f-1aa379f572ae service nova] Releasing lock "refresh_cache-aad1e5ea-5a9c-4499-8f67-2516c5c285cc" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.697156] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquired lock "refresh_cache-aad1e5ea-5a9c-4499-8f67-2516c5c285cc" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.697345] env[61867]: DEBUG nova.network.neutron [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 673.948730] env[61867]: DEBUG nova.compute.utils [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 673.952949] env[61867]: DEBUG nova.compute.manager [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Not allocating networking since 'none' was specified. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 674.082529] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Acquiring lock "f0b6273f-4177-44a3-8dd1-f65faeb07539" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.082631] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Lock "f0b6273f-4177-44a3-8dd1-f65faeb07539" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.218783] env[61867]: DEBUG nova.network.neutron [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.311966] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89c8449f-7add-4d25-a56f-9eae30b4e4d6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.316562] env[61867]: DEBUG nova.network.neutron [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.324019] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13328f4c-fa31-4066-abac-322713460894 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.354282] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21481fd9-3637-4ba4-b795-c697699595d4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.361637] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19499f08-836b-43f5-82c0-fd7b0f2eb1b1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.374879] env[61867]: DEBUG nova.compute.provider_tree [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.454113] env[61867]: DEBUG nova.compute.manager [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 674.630073] env[61867]: DEBUG nova.compute.manager [req-b802d406-fb64-46cb-9f1b-ac76feb29c56 req-13556b7d-33cb-4820-807b-5f8e91fa2cfc service nova] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Received event network-vif-deleted-7ab0af92-a6e0-46c0-bbf5-bda124b94276 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 674.818718] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Releasing lock "refresh_cache-aad1e5ea-5a9c-4499-8f67-2516c5c285cc" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.819128] env[61867]: DEBUG nova.compute.manager [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 674.819316] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 674.819773] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-959bbdb1-a9de-4c92-9435-30eb9325c23c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.828285] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec8c87f-55ad-4a46-bdc5-4bb551b0f39a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.849615] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance aad1e5ea-5a9c-4499-8f67-2516c5c285cc could not be found. [ 674.849931] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 674.850033] env[61867]: INFO nova.compute.manager [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Took 0.03 seconds to destroy the instance on the hypervisor. [ 674.850262] env[61867]: DEBUG oslo.service.loopingcall [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 674.850490] env[61867]: DEBUG nova.compute.manager [-] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 674.850603] env[61867]: DEBUG nova.network.neutron [-] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 674.872747] env[61867]: DEBUG nova.network.neutron [-] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.877528] env[61867]: DEBUG nova.scheduler.client.report [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.378472] env[61867]: DEBUG nova.network.neutron [-] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.381725] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.938s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.382362] env[61867]: ERROR nova.compute.manager [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b80e834f-0219-4832-99f4-94787663a723, please check neutron logs for more information. [ 675.382362] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Traceback (most recent call last): [ 675.382362] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 675.382362] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] self.driver.spawn(context, instance, image_meta, [ 675.382362] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 675.382362] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.382362] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.382362] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] vm_ref = self.build_virtual_machine(instance, [ 675.382362] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.382362] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.382362] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.382878] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] for vif in network_info: [ 675.382878] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 675.382878] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] return self._sync_wrapper(fn, *args, **kwargs) [ 675.382878] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 675.382878] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] self.wait() [ 675.382878] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 675.382878] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] self[:] = self._gt.wait() [ 675.382878] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.382878] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] return self._exit_event.wait() [ 675.382878] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 675.382878] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] current.throw(*self._exc) [ 675.382878] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.382878] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] result = function(*args, **kwargs) [ 675.383294] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.383294] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] return func(*args, **kwargs) [ 675.383294] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.383294] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] raise e [ 675.383294] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.383294] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] nwinfo = self.network_api.allocate_for_instance( [ 675.383294] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.383294] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] created_port_ids = self._update_ports_for_instance( [ 675.383294] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.383294] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] with excutils.save_and_reraise_exception(): [ 675.383294] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.383294] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] self.force_reraise() [ 675.383294] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.383681] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] raise self.value [ 675.383681] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.383681] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] updated_port = self._update_port( [ 675.383681] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.383681] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] _ensure_no_port_binding_failure(port) [ 675.383681] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.383681] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] raise exception.PortBindingFailed(port_id=port['id']) [ 675.383681] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] nova.exception.PortBindingFailed: Binding failed for port b80e834f-0219-4832-99f4-94787663a723, please check neutron logs for more information. [ 675.383681] env[61867]: ERROR nova.compute.manager [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] [ 675.383681] env[61867]: DEBUG nova.compute.utils [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Binding failed for port b80e834f-0219-4832-99f4-94787663a723, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 675.384463] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.515s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.385862] env[61867]: INFO nova.compute.claims [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 675.388719] env[61867]: DEBUG nova.compute.manager [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Build of instance 8f8ea2ec-31fe-4726-8f84-89251c138d44 was re-scheduled: Binding failed for port b80e834f-0219-4832-99f4-94787663a723, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 675.389312] env[61867]: DEBUG nova.compute.manager [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 675.389419] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Acquiring lock "refresh_cache-8f8ea2ec-31fe-4726-8f84-89251c138d44" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.389582] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Acquired lock "refresh_cache-8f8ea2ec-31fe-4726-8f84-89251c138d44" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.389748] env[61867]: DEBUG nova.network.neutron [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 675.466043] env[61867]: DEBUG nova.compute.manager [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 675.489180] env[61867]: DEBUG nova.virt.hardware [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 675.489504] env[61867]: DEBUG nova.virt.hardware [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 675.489689] env[61867]: DEBUG nova.virt.hardware [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 675.489874] env[61867]: DEBUG nova.virt.hardware [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 675.490030] env[61867]: DEBUG nova.virt.hardware [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 675.490179] env[61867]: DEBUG nova.virt.hardware [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 675.490398] env[61867]: DEBUG nova.virt.hardware [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 675.490559] env[61867]: DEBUG nova.virt.hardware [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 675.490734] env[61867]: DEBUG nova.virt.hardware [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 675.490921] env[61867]: DEBUG nova.virt.hardware [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 675.491106] env[61867]: DEBUG nova.virt.hardware [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 675.492228] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5276980a-ed20-4da7-9978-8e531498463e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.500614] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-572315d5-c8fa-4ebd-9b86-b663572d7567 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.514512] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Instance VIF info [] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 675.519772] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Creating folder: Project (c278ea7bd0084d7db9bf1259e84bf69a). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 675.519954] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3678c5ac-2cce-463b-8843-733d25442ee8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.529922] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Created folder: Project (c278ea7bd0084d7db9bf1259e84bf69a) in parent group-v274258. [ 675.530128] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Creating folder: Instances. Parent ref: group-v274275. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 675.530356] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c4cc84ad-87f8-4a60-ba9d-8c10d7503483 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.538066] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Created folder: Instances in parent group-v274275. [ 675.538289] env[61867]: DEBUG oslo.service.loopingcall [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 675.538502] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 675.538697] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6dbd5d0a-a526-4e38-91c6-21746da9e6fd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.555316] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 675.555316] env[61867]: value = "task-1276347" [ 675.555316] env[61867]: _type = "Task" [ 675.555316] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.562886] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276347, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.880945] env[61867]: INFO nova.compute.manager [-] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Took 1.03 seconds to deallocate network for instance. [ 675.885053] env[61867]: DEBUG nova.compute.claims [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 675.885238] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.909484] env[61867]: DEBUG nova.network.neutron [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.001592] env[61867]: DEBUG nova.network.neutron [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.065585] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276347, 'name': CreateVM_Task, 'duration_secs': 0.300407} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.065810] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 676.066652] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.066652] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.066783] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 676.066916] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d924cca-8cb2-45a3-a810-90a8a17caf6f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.071153] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 676.071153] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52426a4e-fbc4-5ac1-f188-aba15b69874d" [ 676.071153] env[61867]: _type = "Task" [ 676.071153] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.078687] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52426a4e-fbc4-5ac1-f188-aba15b69874d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.504410] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Releasing lock "refresh_cache-8f8ea2ec-31fe-4726-8f84-89251c138d44" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.505946] env[61867]: DEBUG nova.compute.manager [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 676.505946] env[61867]: DEBUG nova.compute.manager [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.505946] env[61867]: DEBUG nova.network.neutron [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 676.524121] env[61867]: DEBUG nova.network.neutron [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.583194] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52426a4e-fbc4-5ac1-f188-aba15b69874d, 'name': SearchDatastore_Task, 'duration_secs': 0.026492} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.583499] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.584211] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 676.584211] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.584211] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.584377] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 676.584602] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b3d7a403-1ced-4c08-a508-673e74573ff4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.592245] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 676.592415] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 676.593113] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7f84add-c0cf-4139-8bc9-9aec85bdca95 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.603868] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquiring lock "905ee661-d5bd-4121-9bd0-fe68bdabd2fd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.604163] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "905ee661-d5bd-4121-9bd0-fe68bdabd2fd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.604462] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 676.604462] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5275f4c0-d690-2d0f-4d1d-7a4d390777f3" [ 676.604462] env[61867]: _type = "Task" [ 676.604462] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.619018] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5275f4c0-d690-2d0f-4d1d-7a4d390777f3, 'name': SearchDatastore_Task, 'duration_secs': 0.008087} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.622031] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bed82ae8-2cb6-4430-89c2-1ffc6a1bd4e7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.626801] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 676.626801] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522ae354-ec9f-70fa-615b-d1bd69c8020e" [ 676.626801] env[61867]: _type = "Task" [ 676.626801] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.634975] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522ae354-ec9f-70fa-615b-d1bd69c8020e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.735312] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-739ffdb3-21ce-4416-b9c4-3e59a2a7c117 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.743258] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d877bad-3e45-4a15-b0f7-78258a21f723 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.773529] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56259c9-8305-4ce7-866e-3fe5cfe61307 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.780686] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c10e6d97-d52d-4a5f-af62-cd7b5c88b14a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.793347] env[61867]: DEBUG nova.compute.provider_tree [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.026786] env[61867]: DEBUG nova.network.neutron [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.138031] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522ae354-ec9f-70fa-615b-d1bd69c8020e, 'name': SearchDatastore_Task, 'duration_secs': 0.008689} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.138031] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.138031] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea/37ed9e18-8dba-459d-bc67-e3ce3f9cdaea.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 677.138274] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-459cc596-82ef-4cc8-9cbf-c4f9a7ff9e78 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.144156] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 677.144156] env[61867]: value = "task-1276348" [ 677.144156] env[61867]: _type = "Task" [ 677.144156] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.151419] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276348, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.296915] env[61867]: DEBUG nova.scheduler.client.report [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.529284] env[61867]: INFO nova.compute.manager [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] [instance: 8f8ea2ec-31fe-4726-8f84-89251c138d44] Took 1.02 seconds to deallocate network for instance. [ 677.654339] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276348, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.456997} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.654624] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea/37ed9e18-8dba-459d-bc67-e3ce3f9cdaea.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 677.654835] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 677.655091] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f7eeb099-d641-42cf-bba0-a549b76a2d0e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.661215] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 677.661215] env[61867]: value = "task-1276349" [ 677.661215] env[61867]: _type = "Task" [ 677.661215] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.670273] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276349, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.801670] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.802224] env[61867]: DEBUG nova.compute.manager [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 677.804848] env[61867]: DEBUG oslo_concurrency.lockutils [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.948s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.806342] env[61867]: INFO nova.compute.claims [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 678.171746] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276349, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058968} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.172037] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 678.172795] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c7bff7-7e70-4e0d-927e-9b0d39d58c01 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.191908] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Reconfiguring VM instance instance-00000023 to attach disk [datastore2] 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea/37ed9e18-8dba-459d-bc67-e3ce3f9cdaea.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 678.192208] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40fffc95-598c-4fe2-ad4c-1c7b5b58954f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.212018] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 678.212018] env[61867]: value = "task-1276350" [ 678.212018] env[61867]: _type = "Task" [ 678.212018] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.220052] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276350, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.310767] env[61867]: DEBUG nova.compute.utils [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 678.314095] env[61867]: DEBUG nova.compute.manager [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 678.314273] env[61867]: DEBUG nova.network.neutron [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 678.358035] env[61867]: DEBUG nova.policy [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32f3a7fca4774948a71bdded455edcde', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c8e9d6e6ceac4520add365429c4eb4d2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 678.567985] env[61867]: INFO nova.scheduler.client.report [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Deleted allocations for instance 8f8ea2ec-31fe-4726-8f84-89251c138d44 [ 678.672469] env[61867]: DEBUG nova.network.neutron [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Successfully created port: 5ec8abc2-e15e-4ef5-94d2-3a2c6d0740e4 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 678.723286] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276350, 'name': ReconfigVM_Task, 'duration_secs': 0.253461} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.723802] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Reconfigured VM instance instance-00000023 to attach disk [datastore2] 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea/37ed9e18-8dba-459d-bc67-e3ce3f9cdaea.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 678.726331] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b916bfed-6102-46b4-bb84-c1ef0a4f5b3d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.733257] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 678.733257] env[61867]: value = "task-1276351" [ 678.733257] env[61867]: _type = "Task" [ 678.733257] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.740310] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276351, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.818151] env[61867]: DEBUG nova.compute.manager [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 679.084478] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ca293748-aeb0-4816-acb7-e68f6ae113c1 tempest-ServersTestFqdnHostnames-2065727724 tempest-ServersTestFqdnHostnames-2065727724-project-member] Lock "8f8ea2ec-31fe-4726-8f84-89251c138d44" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.401s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.197109] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e91af0-4da0-488a-bc37-e6a6301489dc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.206866] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6997c1-a4a9-4c60-9586-38e78d469f02 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.241195] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2112f4f8-562c-4125-a343-8f82ed5c2ef9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.250787] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f62f96bb-791e-4694-92b1-5452e1d2d327 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.254267] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276351, 'name': Rename_Task, 'duration_secs': 0.123421} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.254518] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 679.255025] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c9c6b430-4715-4235-91bd-0c665d022938 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.263558] env[61867]: DEBUG nova.compute.provider_tree [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.268839] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 679.268839] env[61867]: value = "task-1276352" [ 679.268839] env[61867]: _type = "Task" [ 679.268839] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.276509] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276352, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.480195] env[61867]: DEBUG nova.compute.manager [req-4e0b3ab2-fc23-48d9-a0d6-f7210da9644a req-53b84ccd-8c90-41e5-8dc8-f2fafcdf5761 service nova] [instance: 5c250675-3708-4c25-b370-6707c457357a] Received event network-changed-5ec8abc2-e15e-4ef5-94d2-3a2c6d0740e4 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 679.480334] env[61867]: DEBUG nova.compute.manager [req-4e0b3ab2-fc23-48d9-a0d6-f7210da9644a req-53b84ccd-8c90-41e5-8dc8-f2fafcdf5761 service nova] [instance: 5c250675-3708-4c25-b370-6707c457357a] Refreshing instance network info cache due to event network-changed-5ec8abc2-e15e-4ef5-94d2-3a2c6d0740e4. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 679.480550] env[61867]: DEBUG oslo_concurrency.lockutils [req-4e0b3ab2-fc23-48d9-a0d6-f7210da9644a req-53b84ccd-8c90-41e5-8dc8-f2fafcdf5761 service nova] Acquiring lock "refresh_cache-5c250675-3708-4c25-b370-6707c457357a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.480695] env[61867]: DEBUG oslo_concurrency.lockutils [req-4e0b3ab2-fc23-48d9-a0d6-f7210da9644a req-53b84ccd-8c90-41e5-8dc8-f2fafcdf5761 service nova] Acquired lock "refresh_cache-5c250675-3708-4c25-b370-6707c457357a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.480923] env[61867]: DEBUG nova.network.neutron [req-4e0b3ab2-fc23-48d9-a0d6-f7210da9644a req-53b84ccd-8c90-41e5-8dc8-f2fafcdf5761 service nova] [instance: 5c250675-3708-4c25-b370-6707c457357a] Refreshing network info cache for port 5ec8abc2-e15e-4ef5-94d2-3a2c6d0740e4 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 679.591240] env[61867]: DEBUG nova.compute.manager [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 679.694663] env[61867]: ERROR nova.compute.manager [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5ec8abc2-e15e-4ef5-94d2-3a2c6d0740e4, please check neutron logs for more information. [ 679.694663] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 679.694663] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.694663] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 679.694663] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.694663] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 679.694663] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.694663] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 679.694663] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.694663] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 679.694663] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.694663] env[61867]: ERROR nova.compute.manager raise self.value [ 679.694663] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.694663] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 679.694663] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.694663] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 679.695248] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.695248] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 679.695248] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5ec8abc2-e15e-4ef5-94d2-3a2c6d0740e4, please check neutron logs for more information. [ 679.695248] env[61867]: ERROR nova.compute.manager [ 679.695248] env[61867]: Traceback (most recent call last): [ 679.695248] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 679.695248] env[61867]: listener.cb(fileno) [ 679.695248] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.695248] env[61867]: result = function(*args, **kwargs) [ 679.695248] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.695248] env[61867]: return func(*args, **kwargs) [ 679.695248] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.695248] env[61867]: raise e [ 679.695248] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.695248] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 679.695248] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.695248] env[61867]: created_port_ids = self._update_ports_for_instance( [ 679.695248] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.695248] env[61867]: with excutils.save_and_reraise_exception(): [ 679.695248] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.695248] env[61867]: self.force_reraise() [ 679.695248] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.695248] env[61867]: raise self.value [ 679.695248] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.695248] env[61867]: updated_port = self._update_port( [ 679.695248] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.695248] env[61867]: _ensure_no_port_binding_failure(port) [ 679.695248] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.695248] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 679.696049] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 5ec8abc2-e15e-4ef5-94d2-3a2c6d0740e4, please check neutron logs for more information. [ 679.696049] env[61867]: Removing descriptor: 17 [ 679.766399] env[61867]: DEBUG nova.scheduler.client.report [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.780843] env[61867]: DEBUG oslo_vmware.api [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276352, 'name': PowerOnVM_Task, 'duration_secs': 0.390545} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.781446] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 679.781446] env[61867]: INFO nova.compute.manager [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Took 4.32 seconds to spawn the instance on the hypervisor. [ 679.781652] env[61867]: DEBUG nova.compute.manager [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 679.782247] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36097c47-fa39-4fcb-9b3f-ea8e22d1a190 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.828180] env[61867]: DEBUG nova.compute.manager [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 679.859083] env[61867]: DEBUG nova.virt.hardware [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 679.859332] env[61867]: DEBUG nova.virt.hardware [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 679.859481] env[61867]: DEBUG nova.virt.hardware [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 679.859730] env[61867]: DEBUG nova.virt.hardware [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 679.859909] env[61867]: DEBUG nova.virt.hardware [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 679.860151] env[61867]: DEBUG nova.virt.hardware [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 679.860417] env[61867]: DEBUG nova.virt.hardware [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 679.860606] env[61867]: DEBUG nova.virt.hardware [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 679.860806] env[61867]: DEBUG nova.virt.hardware [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 679.861009] env[61867]: DEBUG nova.virt.hardware [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 679.861257] env[61867]: DEBUG nova.virt.hardware [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 679.862141] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98bd6d3-5dce-48c8-b8c8-fc38e78cb304 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.871146] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2e27d9-7abe-4d67-8860-158f14ecbf4f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.885415] env[61867]: ERROR nova.compute.manager [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5ec8abc2-e15e-4ef5-94d2-3a2c6d0740e4, please check neutron logs for more information. [ 679.885415] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] Traceback (most recent call last): [ 679.885415] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 679.885415] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] yield resources [ 679.885415] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.885415] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] self.driver.spawn(context, instance, image_meta, [ 679.885415] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 679.885415] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.885415] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.885415] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] vm_ref = self.build_virtual_machine(instance, [ 679.885415] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.885935] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.885935] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.885935] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] for vif in network_info: [ 679.885935] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.885935] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] return self._sync_wrapper(fn, *args, **kwargs) [ 679.885935] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.885935] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] self.wait() [ 679.885935] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.885935] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] self[:] = self._gt.wait() [ 679.885935] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.885935] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] return self._exit_event.wait() [ 679.885935] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 679.885935] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] current.throw(*self._exc) [ 679.887593] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.887593] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] result = function(*args, **kwargs) [ 679.887593] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.887593] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] return func(*args, **kwargs) [ 679.887593] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.887593] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] raise e [ 679.887593] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.887593] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] nwinfo = self.network_api.allocate_for_instance( [ 679.887593] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.887593] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] created_port_ids = self._update_ports_for_instance( [ 679.887593] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.887593] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] with excutils.save_and_reraise_exception(): [ 679.887593] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.888074] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] self.force_reraise() [ 679.888074] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.888074] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] raise self.value [ 679.888074] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.888074] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] updated_port = self._update_port( [ 679.888074] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.888074] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] _ensure_no_port_binding_failure(port) [ 679.888074] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.888074] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] raise exception.PortBindingFailed(port_id=port['id']) [ 679.888074] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] nova.exception.PortBindingFailed: Binding failed for port 5ec8abc2-e15e-4ef5-94d2-3a2c6d0740e4, please check neutron logs for more information. [ 679.888074] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] [ 679.888074] env[61867]: INFO nova.compute.manager [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Terminating instance [ 679.890978] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "refresh_cache-5c250675-3708-4c25-b370-6707c457357a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.001881] env[61867]: DEBUG nova.network.neutron [req-4e0b3ab2-fc23-48d9-a0d6-f7210da9644a req-53b84ccd-8c90-41e5-8dc8-f2fafcdf5761 service nova] [instance: 5c250675-3708-4c25-b370-6707c457357a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.084635] env[61867]: DEBUG nova.network.neutron [req-4e0b3ab2-fc23-48d9-a0d6-f7210da9644a req-53b84ccd-8c90-41e5-8dc8-f2fafcdf5761 service nova] [instance: 5c250675-3708-4c25-b370-6707c457357a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.113523] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.277431] env[61867]: DEBUG oslo_concurrency.lockutils [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.277431] env[61867]: DEBUG nova.compute.manager [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 680.280894] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.364s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.303317] env[61867]: INFO nova.compute.manager [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Took 41.34 seconds to build instance. [ 680.586846] env[61867]: DEBUG oslo_concurrency.lockutils [req-4e0b3ab2-fc23-48d9-a0d6-f7210da9644a req-53b84ccd-8c90-41e5-8dc8-f2fafcdf5761 service nova] Releasing lock "refresh_cache-5c250675-3708-4c25-b370-6707c457357a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.587286] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquired lock "refresh_cache-5c250675-3708-4c25-b370-6707c457357a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.587471] env[61867]: DEBUG nova.network.neutron [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 680.784969] env[61867]: DEBUG nova.compute.utils [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 680.791458] env[61867]: DEBUG nova.compute.manager [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 680.791657] env[61867]: DEBUG nova.network.neutron [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 680.809137] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eebc2269-8eae-4e8a-99b5-9cc8ace2bc38 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Lock "37ed9e18-8dba-459d-bc67-e3ce3f9cdaea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.071s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.854745] env[61867]: DEBUG nova.policy [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e18b96b07e54a1d8ed642b02f11e3a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b64ff57036a492eb2555f33e0af664e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 680.918030] env[61867]: INFO nova.compute.manager [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Rebuilding instance [ 680.965350] env[61867]: DEBUG nova.compute.manager [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 680.967041] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fcc83f9-8519-440c-b6a4-a702845f6fc4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.116017] env[61867]: DEBUG nova.network.neutron [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.243644] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2997b0a3-5948-47d1-b072-302416831344 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.252404] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-461f8e83-ecde-4b03-be18-5c0040803b44 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.257301] env[61867]: DEBUG nova.network.neutron [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.289611] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2b7b518-1aee-44e4-aae4-96cfa3dbfe66 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.292933] env[61867]: DEBUG nova.compute.manager [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 681.297381] env[61867]: DEBUG nova.network.neutron [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Successfully created port: feaf8561-0b7e-4022-ba56-899778e4aaf7 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 681.304725] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-546bd652-d3ce-4eea-a464-3268f107e217 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.319518] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 681.322236] env[61867]: DEBUG nova.compute.provider_tree [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.482679] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 681.482884] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b446aa48-3e10-4605-8175-85485714c4f0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.489440] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 681.489440] env[61867]: value = "task-1276353" [ 681.489440] env[61867]: _type = "Task" [ 681.489440] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.497634] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276353, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.507399] env[61867]: DEBUG nova.compute.manager [req-0c36db82-08b6-452f-b398-db6d78d50c3b req-6b3f6d35-ece0-4090-9c57-39fc6e0c877a service nova] [instance: 5c250675-3708-4c25-b370-6707c457357a] Received event network-vif-deleted-5ec8abc2-e15e-4ef5-94d2-3a2c6d0740e4 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 681.759317] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Releasing lock "refresh_cache-5c250675-3708-4c25-b370-6707c457357a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.759806] env[61867]: DEBUG nova.compute.manager [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 681.759997] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 681.760326] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-59dd2225-edb8-4107-9ea1-fb8ea0f44e2b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.770221] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-013f4b71-c02a-4cc9-8670-5e4440338bcc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.793253] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5c250675-3708-4c25-b370-6707c457357a could not be found. [ 681.793506] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 681.793662] env[61867]: INFO nova.compute.manager [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 681.793929] env[61867]: DEBUG oslo.service.loopingcall [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 681.794141] env[61867]: DEBUG nova.compute.manager [-] [instance: 5c250675-3708-4c25-b370-6707c457357a] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.794672] env[61867]: DEBUG nova.network.neutron [-] [instance: 5c250675-3708-4c25-b370-6707c457357a] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 681.827534] env[61867]: DEBUG nova.network.neutron [-] [instance: 5c250675-3708-4c25-b370-6707c457357a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.829841] env[61867]: DEBUG nova.scheduler.client.report [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.853692] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.004113] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276353, 'name': PowerOffVM_Task, 'duration_secs': 0.209839} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.004113] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 682.004113] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 682.007014] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e061d7d8-fcde-4dbb-94f7-2e00744aac66 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.013750] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 682.013995] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-315587a3-95d1-4204-9ef1-e063af28aa1c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.041467] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 682.041690] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 682.041869] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Deleting the datastore file [datastore2] 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 682.042176] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-149f74e6-2f8e-41b8-a39c-3464b87a3080 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.048436] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 682.048436] env[61867]: value = "task-1276355" [ 682.048436] env[61867]: _type = "Task" [ 682.048436] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.055955] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276355, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.304743] env[61867]: DEBUG nova.compute.manager [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 682.335702] env[61867]: DEBUG nova.virt.hardware [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 682.335936] env[61867]: DEBUG nova.virt.hardware [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 682.336083] env[61867]: DEBUG nova.virt.hardware [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 682.336262] env[61867]: DEBUG nova.virt.hardware [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 682.336405] env[61867]: DEBUG nova.virt.hardware [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 682.336574] env[61867]: DEBUG nova.virt.hardware [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 682.336792] env[61867]: DEBUG nova.virt.hardware [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 682.336947] env[61867]: DEBUG nova.virt.hardware [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 682.340023] env[61867]: DEBUG nova.virt.hardware [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 682.340023] env[61867]: DEBUG nova.virt.hardware [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 682.340023] env[61867]: DEBUG nova.virt.hardware [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 682.340023] env[61867]: DEBUG nova.network.neutron [-] [instance: 5c250675-3708-4c25-b370-6707c457357a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.340023] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.059s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.340328] env[61867]: ERROR nova.compute.manager [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 91bb10e6-7994-433e-a261-e3d0ae9c2d72, please check neutron logs for more information. [ 682.340328] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Traceback (most recent call last): [ 682.340328] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 682.340328] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] self.driver.spawn(context, instance, image_meta, [ 682.340328] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 682.340328] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.340328] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.340328] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] vm_ref = self.build_virtual_machine(instance, [ 682.340328] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.340328] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.340328] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.340602] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] for vif in network_info: [ 682.340602] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 682.340602] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] return self._sync_wrapper(fn, *args, **kwargs) [ 682.340602] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 682.340602] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] self.wait() [ 682.340602] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 682.340602] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] self[:] = self._gt.wait() [ 682.340602] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.340602] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] return self._exit_event.wait() [ 682.340602] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 682.340602] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] current.throw(*self._exc) [ 682.340602] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.340602] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] result = function(*args, **kwargs) [ 682.340917] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 682.340917] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] return func(*args, **kwargs) [ 682.340917] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.340917] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] raise e [ 682.340917] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.340917] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] nwinfo = self.network_api.allocate_for_instance( [ 682.340917] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.340917] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] created_port_ids = self._update_ports_for_instance( [ 682.340917] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.340917] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] with excutils.save_and_reraise_exception(): [ 682.340917] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.340917] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] self.force_reraise() [ 682.340917] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.341373] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] raise self.value [ 682.341373] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.341373] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] updated_port = self._update_port( [ 682.341373] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.341373] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] _ensure_no_port_binding_failure(port) [ 682.341373] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.341373] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] raise exception.PortBindingFailed(port_id=port['id']) [ 682.341373] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] nova.exception.PortBindingFailed: Binding failed for port 91bb10e6-7994-433e-a261-e3d0ae9c2d72, please check neutron logs for more information. [ 682.341373] env[61867]: ERROR nova.compute.manager [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] [ 682.341373] env[61867]: DEBUG nova.compute.utils [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Binding failed for port 91bb10e6-7994-433e-a261-e3d0ae9c2d72, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 682.344599] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-522dd9c0-387c-4506-9f9c-a641b747c5fa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.345314] env[61867]: DEBUG nova.compute.manager [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Build of instance 50930c54-d03e-4529-ac85-a73e9a5b4e4d was re-scheduled: Binding failed for port 91bb10e6-7994-433e-a261-e3d0ae9c2d72, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 682.345648] env[61867]: DEBUG nova.compute.manager [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 682.345971] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Acquiring lock "refresh_cache-50930c54-d03e-4529-ac85-a73e9a5b4e4d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.346073] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Acquired lock "refresh_cache-50930c54-d03e-4529-ac85-a73e9a5b4e4d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.346164] env[61867]: DEBUG nova.network.neutron [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 682.347120] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.007s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.355558] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e14fc0e-aa7a-40db-8887-c54980d3f37a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.433807] env[61867]: ERROR nova.compute.manager [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port feaf8561-0b7e-4022-ba56-899778e4aaf7, please check neutron logs for more information. [ 682.433807] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 682.433807] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.433807] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 682.433807] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.433807] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 682.433807] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.433807] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 682.433807] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.433807] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 682.433807] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.433807] env[61867]: ERROR nova.compute.manager raise self.value [ 682.433807] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.433807] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 682.433807] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.433807] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 682.434569] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.434569] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 682.434569] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port feaf8561-0b7e-4022-ba56-899778e4aaf7, please check neutron logs for more information. [ 682.434569] env[61867]: ERROR nova.compute.manager [ 682.434569] env[61867]: Traceback (most recent call last): [ 682.434569] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 682.434569] env[61867]: listener.cb(fileno) [ 682.434569] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.434569] env[61867]: result = function(*args, **kwargs) [ 682.434569] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 682.434569] env[61867]: return func(*args, **kwargs) [ 682.434569] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.434569] env[61867]: raise e [ 682.434569] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.434569] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 682.434569] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.434569] env[61867]: created_port_ids = self._update_ports_for_instance( [ 682.434569] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.434569] env[61867]: with excutils.save_and_reraise_exception(): [ 682.434569] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.434569] env[61867]: self.force_reraise() [ 682.434569] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.434569] env[61867]: raise self.value [ 682.434569] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.434569] env[61867]: updated_port = self._update_port( [ 682.434569] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.434569] env[61867]: _ensure_no_port_binding_failure(port) [ 682.434569] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.434569] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 682.435715] env[61867]: nova.exception.PortBindingFailed: Binding failed for port feaf8561-0b7e-4022-ba56-899778e4aaf7, please check neutron logs for more information. [ 682.435715] env[61867]: Removing descriptor: 17 [ 682.435715] env[61867]: ERROR nova.compute.manager [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port feaf8561-0b7e-4022-ba56-899778e4aaf7, please check neutron logs for more information. [ 682.435715] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Traceback (most recent call last): [ 682.435715] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 682.435715] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] yield resources [ 682.435715] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 682.435715] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] self.driver.spawn(context, instance, image_meta, [ 682.435715] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 682.435715] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.435715] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.435715] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] vm_ref = self.build_virtual_machine(instance, [ 682.436220] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.436220] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.436220] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.436220] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] for vif in network_info: [ 682.436220] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 682.436220] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] return self._sync_wrapper(fn, *args, **kwargs) [ 682.436220] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 682.436220] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] self.wait() [ 682.436220] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 682.436220] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] self[:] = self._gt.wait() [ 682.436220] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.436220] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] return self._exit_event.wait() [ 682.436220] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 682.436685] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] result = hub.switch() [ 682.436685] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 682.436685] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] return self.greenlet.switch() [ 682.436685] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.436685] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] result = function(*args, **kwargs) [ 682.436685] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 682.436685] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] return func(*args, **kwargs) [ 682.436685] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.436685] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] raise e [ 682.436685] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.436685] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] nwinfo = self.network_api.allocate_for_instance( [ 682.436685] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.436685] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] created_port_ids = self._update_ports_for_instance( [ 682.437352] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.437352] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] with excutils.save_and_reraise_exception(): [ 682.437352] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.437352] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] self.force_reraise() [ 682.437352] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.437352] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] raise self.value [ 682.437352] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.437352] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] updated_port = self._update_port( [ 682.437352] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.437352] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] _ensure_no_port_binding_failure(port) [ 682.437352] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.437352] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] raise exception.PortBindingFailed(port_id=port['id']) [ 682.437939] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] nova.exception.PortBindingFailed: Binding failed for port feaf8561-0b7e-4022-ba56-899778e4aaf7, please check neutron logs for more information. [ 682.437939] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] [ 682.437939] env[61867]: INFO nova.compute.manager [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Terminating instance [ 682.437939] env[61867]: DEBUG oslo_concurrency.lockutils [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "refresh_cache-4d9c1c85-9d49-4c7c-bcac-f22f5842db0d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.437939] env[61867]: DEBUG oslo_concurrency.lockutils [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquired lock "refresh_cache-4d9c1c85-9d49-4c7c-bcac-f22f5842db0d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.437939] env[61867]: DEBUG nova.network.neutron [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 682.558265] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276355, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.099982} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.558467] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 682.558696] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 682.558875] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 682.845880] env[61867]: INFO nova.compute.manager [-] [instance: 5c250675-3708-4c25-b370-6707c457357a] Took 1.05 seconds to deallocate network for instance. [ 682.852364] env[61867]: DEBUG nova.compute.claims [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 682.852545] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.868209] env[61867]: DEBUG nova.network.neutron [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.957453] env[61867]: DEBUG nova.network.neutron [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.963282] env[61867]: DEBUG nova.network.neutron [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.005438] env[61867]: DEBUG nova.network.neutron [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.187127] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22892423-7589-4771-ac5f-2b45ae167706 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.195804] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db30d10-2f15-4d2e-8aa0-a276d64c7594 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.225027] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5274179a-5413-4e8c-825d-2c0b8e1b4b2f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.232293] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa9322b-f35b-4a10-a8fc-5663781b8bfa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.245550] env[61867]: DEBUG nova.compute.provider_tree [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.467751] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Releasing lock "refresh_cache-50930c54-d03e-4529-ac85-a73e9a5b4e4d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.468030] env[61867]: DEBUG nova.compute.manager [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 683.468223] env[61867]: DEBUG nova.compute.manager [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 683.468394] env[61867]: DEBUG nova.network.neutron [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.483022] env[61867]: DEBUG nova.network.neutron [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.507514] env[61867]: DEBUG oslo_concurrency.lockutils [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Releasing lock "refresh_cache-4d9c1c85-9d49-4c7c-bcac-f22f5842db0d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.507914] env[61867]: DEBUG nova.compute.manager [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 683.508117] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 683.508416] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d7baa1dc-099d-4496-a8f8-06e504899589 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.517590] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082bbba2-9854-4c13-b7d7-ec2e5da2871c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.531045] env[61867]: DEBUG nova.compute.manager [req-ed10a46b-2031-48df-8a6e-8de32595fcec req-1836f48a-71c3-4029-8b0d-00420dbe484d service nova] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Received event network-changed-feaf8561-0b7e-4022-ba56-899778e4aaf7 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 683.531045] env[61867]: DEBUG nova.compute.manager [req-ed10a46b-2031-48df-8a6e-8de32595fcec req-1836f48a-71c3-4029-8b0d-00420dbe484d service nova] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Refreshing instance network info cache due to event network-changed-feaf8561-0b7e-4022-ba56-899778e4aaf7. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 683.531045] env[61867]: DEBUG oslo_concurrency.lockutils [req-ed10a46b-2031-48df-8a6e-8de32595fcec req-1836f48a-71c3-4029-8b0d-00420dbe484d service nova] Acquiring lock "refresh_cache-4d9c1c85-9d49-4c7c-bcac-f22f5842db0d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.531424] env[61867]: DEBUG oslo_concurrency.lockutils [req-ed10a46b-2031-48df-8a6e-8de32595fcec req-1836f48a-71c3-4029-8b0d-00420dbe484d service nova] Acquired lock "refresh_cache-4d9c1c85-9d49-4c7c-bcac-f22f5842db0d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.531424] env[61867]: DEBUG nova.network.neutron [req-ed10a46b-2031-48df-8a6e-8de32595fcec req-1836f48a-71c3-4029-8b0d-00420dbe484d service nova] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Refreshing network info cache for port feaf8561-0b7e-4022-ba56-899778e4aaf7 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 683.540579] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d could not be found. [ 683.540790] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 683.540963] env[61867]: INFO nova.compute.manager [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 683.541206] env[61867]: DEBUG oslo.service.loopingcall [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 683.541962] env[61867]: DEBUG nova.compute.manager [-] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 683.542072] env[61867]: DEBUG nova.network.neutron [-] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.556457] env[61867]: DEBUG nova.network.neutron [-] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.591953] env[61867]: DEBUG nova.virt.hardware [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 683.592217] env[61867]: DEBUG nova.virt.hardware [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 683.592372] env[61867]: DEBUG nova.virt.hardware [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 683.592549] env[61867]: DEBUG nova.virt.hardware [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 683.592695] env[61867]: DEBUG nova.virt.hardware [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 683.592840] env[61867]: DEBUG nova.virt.hardware [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 683.593055] env[61867]: DEBUG nova.virt.hardware [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 683.593221] env[61867]: DEBUG nova.virt.hardware [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 683.593382] env[61867]: DEBUG nova.virt.hardware [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 683.593541] env[61867]: DEBUG nova.virt.hardware [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 683.593711] env[61867]: DEBUG nova.virt.hardware [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 683.594532] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f5c2bb-ba45-4dfc-ad37-ece5d0a121e5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.602778] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40a054e-50e6-4d1d-8f1d-f282f2ce0842 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.615655] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Instance VIF info [] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 683.620998] env[61867]: DEBUG oslo.service.loopingcall [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 683.621227] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 683.621410] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a8cb7c49-412a-430f-ba69-a02ab465dbcf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.637390] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 683.637390] env[61867]: value = "task-1276356" [ 683.637390] env[61867]: _type = "Task" [ 683.637390] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.644084] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276356, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.748074] env[61867]: DEBUG nova.scheduler.client.report [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 683.985791] env[61867]: DEBUG nova.network.neutron [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.051159] env[61867]: DEBUG nova.network.neutron [req-ed10a46b-2031-48df-8a6e-8de32595fcec req-1836f48a-71c3-4029-8b0d-00420dbe484d service nova] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.057926] env[61867]: DEBUG nova.network.neutron [-] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.104414] env[61867]: DEBUG nova.network.neutron [req-ed10a46b-2031-48df-8a6e-8de32595fcec req-1836f48a-71c3-4029-8b0d-00420dbe484d service nova] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.147754] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276356, 'name': CreateVM_Task, 'duration_secs': 0.254352} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.147877] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 684.148297] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.148454] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.148862] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 684.149053] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4eb05759-644d-4c36-9818-6d7a438ddb85 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.153577] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 684.153577] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525c0b61-800f-70c0-7f08-5897fd174d3f" [ 684.153577] env[61867]: _type = "Task" [ 684.153577] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.160992] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525c0b61-800f-70c0-7f08-5897fd174d3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.252967] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.906s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.253667] env[61867]: ERROR nova.compute.manager [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 783c055f-622e-4107-b61b-e923222aa69c, please check neutron logs for more information. [ 684.253667] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Traceback (most recent call last): [ 684.253667] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.253667] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] self.driver.spawn(context, instance, image_meta, [ 684.253667] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 684.253667] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.253667] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.253667] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] vm_ref = self.build_virtual_machine(instance, [ 684.253667] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.253667] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.253667] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.253949] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] for vif in network_info: [ 684.253949] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.253949] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] return self._sync_wrapper(fn, *args, **kwargs) [ 684.253949] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.253949] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] self.wait() [ 684.253949] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.253949] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] self[:] = self._gt.wait() [ 684.253949] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.253949] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] return self._exit_event.wait() [ 684.253949] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 684.253949] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] current.throw(*self._exc) [ 684.253949] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.253949] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] result = function(*args, **kwargs) [ 684.254326] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 684.254326] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] return func(*args, **kwargs) [ 684.254326] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.254326] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] raise e [ 684.254326] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.254326] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] nwinfo = self.network_api.allocate_for_instance( [ 684.254326] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.254326] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] created_port_ids = self._update_ports_for_instance( [ 684.254326] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.254326] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] with excutils.save_and_reraise_exception(): [ 684.254326] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.254326] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] self.force_reraise() [ 684.254326] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.254642] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] raise self.value [ 684.254642] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.254642] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] updated_port = self._update_port( [ 684.254642] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.254642] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] _ensure_no_port_binding_failure(port) [ 684.254642] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.254642] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] raise exception.PortBindingFailed(port_id=port['id']) [ 684.254642] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] nova.exception.PortBindingFailed: Binding failed for port 783c055f-622e-4107-b61b-e923222aa69c, please check neutron logs for more information. [ 684.254642] env[61867]: ERROR nova.compute.manager [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] [ 684.254642] env[61867]: DEBUG nova.compute.utils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Binding failed for port 783c055f-622e-4107-b61b-e923222aa69c, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 684.255531] env[61867]: DEBUG oslo_concurrency.lockutils [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.444s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.257051] env[61867]: INFO nova.compute.claims [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 684.260778] env[61867]: DEBUG nova.compute.manager [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Build of instance 572dbc25-39a6-4e8b-b5c0-176ea0e27839 was re-scheduled: Binding failed for port 783c055f-622e-4107-b61b-e923222aa69c, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 684.261225] env[61867]: DEBUG nova.compute.manager [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 684.261454] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Acquiring lock "refresh_cache-572dbc25-39a6-4e8b-b5c0-176ea0e27839" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.261602] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Acquired lock "refresh_cache-572dbc25-39a6-4e8b-b5c0-176ea0e27839" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.261759] env[61867]: DEBUG nova.network.neutron [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 684.488850] env[61867]: INFO nova.compute.manager [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] [instance: 50930c54-d03e-4529-ac85-a73e9a5b4e4d] Took 1.02 seconds to deallocate network for instance. [ 684.560224] env[61867]: INFO nova.compute.manager [-] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Took 1.02 seconds to deallocate network for instance. [ 684.562373] env[61867]: DEBUG nova.compute.claims [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 684.562551] env[61867]: DEBUG oslo_concurrency.lockutils [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.606987] env[61867]: DEBUG oslo_concurrency.lockutils [req-ed10a46b-2031-48df-8a6e-8de32595fcec req-1836f48a-71c3-4029-8b0d-00420dbe484d service nova] Releasing lock "refresh_cache-4d9c1c85-9d49-4c7c-bcac-f22f5842db0d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.607820] env[61867]: DEBUG nova.compute.manager [req-ed10a46b-2031-48df-8a6e-8de32595fcec req-1836f48a-71c3-4029-8b0d-00420dbe484d service nova] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Received event network-vif-deleted-feaf8561-0b7e-4022-ba56-899778e4aaf7 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 684.664886] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525c0b61-800f-70c0-7f08-5897fd174d3f, 'name': SearchDatastore_Task, 'duration_secs': 0.008281} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.665191] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.665408] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 684.665628] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.665773] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.665942] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 684.666202] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1f861f51-4459-4791-894b-50049990ba82 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.674236] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 684.674401] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 684.675100] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ef88cad-22c6-4a9e-b581-a9e26c611e30 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.679740] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 684.679740] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52616441-e2c4-78ca-dd44-ab1f4262e380" [ 684.679740] env[61867]: _type = "Task" [ 684.679740] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.686810] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52616441-e2c4-78ca-dd44-ab1f4262e380, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.778184] env[61867]: DEBUG nova.network.neutron [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.847964] env[61867]: DEBUG nova.network.neutron [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.190812] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52616441-e2c4-78ca-dd44-ab1f4262e380, 'name': SearchDatastore_Task, 'duration_secs': 0.012321} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.191666] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80838f94-1cc6-4284-8f5c-9312b980d6bb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.197283] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 685.197283] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52a798a3-ed40-4f4b-ddf1-a9f52692b716" [ 685.197283] env[61867]: _type = "Task" [ 685.197283] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.203867] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52a798a3-ed40-4f4b-ddf1-a9f52692b716, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.352937] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Releasing lock "refresh_cache-572dbc25-39a6-4e8b-b5c0-176ea0e27839" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.352937] env[61867]: DEBUG nova.compute.manager [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 685.353082] env[61867]: DEBUG nova.compute.manager [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 685.353180] env[61867]: DEBUG nova.network.neutron [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 685.371035] env[61867]: DEBUG nova.network.neutron [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.522844] env[61867]: INFO nova.scheduler.client.report [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Deleted allocations for instance 50930c54-d03e-4529-ac85-a73e9a5b4e4d [ 685.609737] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba630d70-55db-4b1c-bf73-3a17b30d2625 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.619674] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a00514-ebbe-4ca9-bb06-310fc52c33bd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.668064] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-806547c8-5223-4182-a9e9-70b6c41c64dd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.677721] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0fd402d-20e2-4c01-bca7-7da21cff5ad0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.691453] env[61867]: DEBUG nova.compute.provider_tree [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.705552] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52a798a3-ed40-4f4b-ddf1-a9f52692b716, 'name': SearchDatastore_Task, 'duration_secs': 0.008752} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.705872] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.706085] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea/37ed9e18-8dba-459d-bc67-e3ce3f9cdaea.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 685.706343] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3c2ce6f6-71d0-423a-85bc-f069762ce9c6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.712599] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 685.712599] env[61867]: value = "task-1276357" [ 685.712599] env[61867]: _type = "Task" [ 685.712599] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.720121] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276357, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.874064] env[61867]: DEBUG nova.network.neutron [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.042887] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5eebac3b-5f7e-40f5-ba4a-22b48061388a tempest-ServersV294TestFqdnHostnames-286502690 tempest-ServersV294TestFqdnHostnames-286502690-project-member] Lock "50930c54-d03e-4529-ac85-a73e9a5b4e4d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.543s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.194183] env[61867]: DEBUG nova.scheduler.client.report [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 686.225292] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276357, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.452023} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.225550] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea/37ed9e18-8dba-459d-bc67-e3ce3f9cdaea.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 686.225763] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 686.226011] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-57f75479-819b-493a-ab67-18a056b602a7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.232971] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 686.232971] env[61867]: value = "task-1276358" [ 686.232971] env[61867]: _type = "Task" [ 686.232971] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.241195] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276358, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.376769] env[61867]: INFO nova.compute.manager [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] [instance: 572dbc25-39a6-4e8b-b5c0-176ea0e27839] Took 1.02 seconds to deallocate network for instance. [ 686.545386] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 686.699114] env[61867]: DEBUG oslo_concurrency.lockutils [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.699684] env[61867]: DEBUG nova.compute.manager [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 686.702160] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.399s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.744056] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276358, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057617} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.744372] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 686.745021] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac84849-5527-4d40-aed2-f4f5cf595a2a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.765582] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Reconfiguring VM instance instance-00000023 to attach disk [datastore2] 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea/37ed9e18-8dba-459d-bc67-e3ce3f9cdaea.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 686.765800] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-016b069e-8985-41c7-bd85-6e33ed9d0a0f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.785403] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 686.785403] env[61867]: value = "task-1276359" [ 686.785403] env[61867]: _type = "Task" [ 686.785403] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.793075] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276359, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.071887] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.207725] env[61867]: DEBUG nova.compute.utils [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 687.213455] env[61867]: DEBUG nova.compute.manager [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 687.213687] env[61867]: DEBUG nova.network.neutron [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 687.262962] env[61867]: DEBUG nova.policy [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a7f202aa2b24bae8133ed4378d88d35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ca75e7b9e68c4d7597cf8126c9edd7db', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 687.296579] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276359, 'name': ReconfigVM_Task, 'duration_secs': 0.278174} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.297259] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Reconfigured VM instance instance-00000023 to attach disk [datastore2] 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea/37ed9e18-8dba-459d-bc67-e3ce3f9cdaea.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 687.297873] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b8ec04ab-fcbe-429a-bf77-7a29bb647432 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.303859] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 687.303859] env[61867]: value = "task-1276360" [ 687.303859] env[61867]: _type = "Task" [ 687.303859] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.314016] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276360, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.412925] env[61867]: INFO nova.scheduler.client.report [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Deleted allocations for instance 572dbc25-39a6-4e8b-b5c0-176ea0e27839 [ 687.549853] env[61867]: DEBUG nova.network.neutron [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Successfully created port: 78d19f74-61d7-47cc-bafd-33d2683cb241 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 687.601801] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cffbeea0-e002-4265-a2cc-b1b18534f3f0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.609345] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aada943a-27f0-4139-a0ac-a82a6e85074e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.640259] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b9c6b04-8ddf-4f3a-adf2-0569efc10d4a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.647822] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efddb0b7-3b95-4b0f-8eb5-e25dd1d6b36e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.662745] env[61867]: DEBUG nova.compute.provider_tree [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.717145] env[61867]: DEBUG nova.compute.manager [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 687.813871] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276360, 'name': Rename_Task, 'duration_secs': 0.150466} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.817252] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 687.817495] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d10c283-2eb2-4afe-b174-15b74148da1c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.824807] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 687.824807] env[61867]: value = "task-1276361" [ 687.824807] env[61867]: _type = "Task" [ 687.824807] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.837859] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276361, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.926395] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b504ed37-2116-4e1f-a56d-cfc94bf89092 tempest-ServerGroupTestJSON-605040800 tempest-ServerGroupTestJSON-605040800-project-member] Lock "572dbc25-39a6-4e8b-b5c0-176ea0e27839" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.032s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.165368] env[61867]: DEBUG nova.scheduler.client.report [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 688.309110] env[61867]: DEBUG nova.compute.manager [req-7fa6d458-d0c2-4025-b696-c4e3ed7ee892 req-32ef9646-bcbf-4c89-9c4f-a4ba4d7860b8 service nova] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Received event network-changed-78d19f74-61d7-47cc-bafd-33d2683cb241 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 688.309826] env[61867]: DEBUG nova.compute.manager [req-7fa6d458-d0c2-4025-b696-c4e3ed7ee892 req-32ef9646-bcbf-4c89-9c4f-a4ba4d7860b8 service nova] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Refreshing instance network info cache due to event network-changed-78d19f74-61d7-47cc-bafd-33d2683cb241. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 688.310435] env[61867]: DEBUG oslo_concurrency.lockutils [req-7fa6d458-d0c2-4025-b696-c4e3ed7ee892 req-32ef9646-bcbf-4c89-9c4f-a4ba4d7860b8 service nova] Acquiring lock "refresh_cache-59d2ee3a-756f-4dda-a70a-6d43aaf32049" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.310947] env[61867]: DEBUG oslo_concurrency.lockutils [req-7fa6d458-d0c2-4025-b696-c4e3ed7ee892 req-32ef9646-bcbf-4c89-9c4f-a4ba4d7860b8 service nova] Acquired lock "refresh_cache-59d2ee3a-756f-4dda-a70a-6d43aaf32049" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.311771] env[61867]: DEBUG nova.network.neutron [req-7fa6d458-d0c2-4025-b696-c4e3ed7ee892 req-32ef9646-bcbf-4c89-9c4f-a4ba4d7860b8 service nova] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Refreshing network info cache for port 78d19f74-61d7-47cc-bafd-33d2683cb241 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 688.337928] env[61867]: DEBUG oslo_vmware.api [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276361, 'name': PowerOnVM_Task, 'duration_secs': 0.416022} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.338439] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 688.339035] env[61867]: DEBUG nova.compute.manager [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 688.340512] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5970048e-13bf-4ff3-ba20-53d8a9f8b3b1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.428736] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 688.462533] env[61867]: ERROR nova.compute.manager [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 78d19f74-61d7-47cc-bafd-33d2683cb241, please check neutron logs for more information. [ 688.462533] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 688.462533] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.462533] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 688.462533] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.462533] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 688.462533] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.462533] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 688.462533] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.462533] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 688.462533] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.462533] env[61867]: ERROR nova.compute.manager raise self.value [ 688.462533] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.462533] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 688.462533] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.462533] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 688.463066] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.463066] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 688.463066] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 78d19f74-61d7-47cc-bafd-33d2683cb241, please check neutron logs for more information. [ 688.463066] env[61867]: ERROR nova.compute.manager [ 688.463066] env[61867]: Traceback (most recent call last): [ 688.463066] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 688.463066] env[61867]: listener.cb(fileno) [ 688.463066] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.463066] env[61867]: result = function(*args, **kwargs) [ 688.463066] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 688.463066] env[61867]: return func(*args, **kwargs) [ 688.463066] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 688.463066] env[61867]: raise e [ 688.463066] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.463066] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 688.463066] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.463066] env[61867]: created_port_ids = self._update_ports_for_instance( [ 688.463066] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.463066] env[61867]: with excutils.save_and_reraise_exception(): [ 688.463066] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.463066] env[61867]: self.force_reraise() [ 688.463066] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.463066] env[61867]: raise self.value [ 688.463066] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.463066] env[61867]: updated_port = self._update_port( [ 688.463066] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.463066] env[61867]: _ensure_no_port_binding_failure(port) [ 688.463066] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.463066] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 688.464030] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 78d19f74-61d7-47cc-bafd-33d2683cb241, please check neutron logs for more information. [ 688.464030] env[61867]: Removing descriptor: 21 [ 688.674181] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.972s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.674803] env[61867]: ERROR nova.compute.manager [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 556c9e4c-9a3a-4e1f-9c99-ea0e3331ef02, please check neutron logs for more information. [ 688.674803] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Traceback (most recent call last): [ 688.674803] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 688.674803] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] self.driver.spawn(context, instance, image_meta, [ 688.674803] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 688.674803] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 688.674803] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 688.674803] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] vm_ref = self.build_virtual_machine(instance, [ 688.674803] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 688.674803] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] vif_infos = vmwarevif.get_vif_info(self._session, [ 688.674803] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 688.675171] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] for vif in network_info: [ 688.675171] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 688.675171] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] return self._sync_wrapper(fn, *args, **kwargs) [ 688.675171] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 688.675171] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] self.wait() [ 688.675171] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 688.675171] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] self[:] = self._gt.wait() [ 688.675171] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 688.675171] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] return self._exit_event.wait() [ 688.675171] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 688.675171] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] current.throw(*self._exc) [ 688.675171] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.675171] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] result = function(*args, **kwargs) [ 688.675550] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 688.675550] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] return func(*args, **kwargs) [ 688.675550] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 688.675550] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] raise e [ 688.675550] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.675550] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] nwinfo = self.network_api.allocate_for_instance( [ 688.675550] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.675550] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] created_port_ids = self._update_ports_for_instance( [ 688.675550] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.675550] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] with excutils.save_and_reraise_exception(): [ 688.675550] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.675550] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] self.force_reraise() [ 688.675550] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.675940] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] raise self.value [ 688.675940] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.675940] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] updated_port = self._update_port( [ 688.675940] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.675940] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] _ensure_no_port_binding_failure(port) [ 688.675940] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.675940] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] raise exception.PortBindingFailed(port_id=port['id']) [ 688.675940] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] nova.exception.PortBindingFailed: Binding failed for port 556c9e4c-9a3a-4e1f-9c99-ea0e3331ef02, please check neutron logs for more information. [ 688.675940] env[61867]: ERROR nova.compute.manager [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] [ 688.675940] env[61867]: DEBUG nova.compute.utils [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Binding failed for port 556c9e4c-9a3a-4e1f-9c99-ea0e3331ef02, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 688.676738] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.463s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.678241] env[61867]: INFO nova.compute.claims [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 688.680890] env[61867]: DEBUG nova.compute.manager [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Build of instance 274c5625-eb81-45d1-bd95-7336bddfad1e was re-scheduled: Binding failed for port 556c9e4c-9a3a-4e1f-9c99-ea0e3331ef02, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 688.681343] env[61867]: DEBUG nova.compute.manager [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 688.681609] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "refresh_cache-274c5625-eb81-45d1-bd95-7336bddfad1e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.681765] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "refresh_cache-274c5625-eb81-45d1-bd95-7336bddfad1e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.681950] env[61867]: DEBUG nova.network.neutron [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 688.730908] env[61867]: DEBUG nova.compute.manager [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 688.757976] env[61867]: DEBUG nova.virt.hardware [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 688.759071] env[61867]: DEBUG nova.virt.hardware [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 688.759071] env[61867]: DEBUG nova.virt.hardware [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 688.759071] env[61867]: DEBUG nova.virt.hardware [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 688.759071] env[61867]: DEBUG nova.virt.hardware [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 688.759271] env[61867]: DEBUG nova.virt.hardware [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 688.759306] env[61867]: DEBUG nova.virt.hardware [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 688.759462] env[61867]: DEBUG nova.virt.hardware [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 688.761116] env[61867]: DEBUG nova.virt.hardware [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 688.761116] env[61867]: DEBUG nova.virt.hardware [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 688.761116] env[61867]: DEBUG nova.virt.hardware [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 688.761116] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64188ed2-4ef7-452b-89cb-12130587be5a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.771878] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bda1943-cd99-4739-add6-3884785870ff {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.786152] env[61867]: ERROR nova.compute.manager [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 78d19f74-61d7-47cc-bafd-33d2683cb241, please check neutron logs for more information. [ 688.786152] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Traceback (most recent call last): [ 688.786152] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 688.786152] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] yield resources [ 688.786152] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 688.786152] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] self.driver.spawn(context, instance, image_meta, [ 688.786152] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 688.786152] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] self._vmops.spawn(context, instance, image_meta, injected_files, [ 688.786152] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 688.786152] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] vm_ref = self.build_virtual_machine(instance, [ 688.786152] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 688.786505] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] vif_infos = vmwarevif.get_vif_info(self._session, [ 688.786505] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 688.786505] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] for vif in network_info: [ 688.786505] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 688.786505] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] return self._sync_wrapper(fn, *args, **kwargs) [ 688.786505] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 688.786505] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] self.wait() [ 688.786505] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 688.786505] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] self[:] = self._gt.wait() [ 688.786505] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 688.786505] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] return self._exit_event.wait() [ 688.786505] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 688.786505] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] current.throw(*self._exc) [ 688.786965] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.786965] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] result = function(*args, **kwargs) [ 688.786965] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 688.786965] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] return func(*args, **kwargs) [ 688.786965] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 688.786965] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] raise e [ 688.786965] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.786965] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] nwinfo = self.network_api.allocate_for_instance( [ 688.786965] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.786965] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] created_port_ids = self._update_ports_for_instance( [ 688.786965] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.786965] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] with excutils.save_and_reraise_exception(): [ 688.786965] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.787307] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] self.force_reraise() [ 688.787307] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.787307] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] raise self.value [ 688.787307] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.787307] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] updated_port = self._update_port( [ 688.787307] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.787307] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] _ensure_no_port_binding_failure(port) [ 688.787307] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.787307] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] raise exception.PortBindingFailed(port_id=port['id']) [ 688.787307] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] nova.exception.PortBindingFailed: Binding failed for port 78d19f74-61d7-47cc-bafd-33d2683cb241, please check neutron logs for more information. [ 688.787307] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] [ 688.787307] env[61867]: INFO nova.compute.manager [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Terminating instance [ 688.788532] env[61867]: DEBUG oslo_concurrency.lockutils [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Acquiring lock "refresh_cache-59d2ee3a-756f-4dda-a70a-6d43aaf32049" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.833653] env[61867]: DEBUG nova.network.neutron [req-7fa6d458-d0c2-4025-b696-c4e3ed7ee892 req-32ef9646-bcbf-4c89-9c4f-a4ba4d7860b8 service nova] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.862617] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.952579] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.969449] env[61867]: DEBUG nova.network.neutron [req-7fa6d458-d0c2-4025-b696-c4e3ed7ee892 req-32ef9646-bcbf-4c89-9c4f-a4ba4d7860b8 service nova] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.210342] env[61867]: DEBUG nova.network.neutron [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.342250] env[61867]: DEBUG nova.network.neutron [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.472284] env[61867]: DEBUG oslo_concurrency.lockutils [req-7fa6d458-d0c2-4025-b696-c4e3ed7ee892 req-32ef9646-bcbf-4c89-9c4f-a4ba4d7860b8 service nova] Releasing lock "refresh_cache-59d2ee3a-756f-4dda-a70a-6d43aaf32049" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.472605] env[61867]: DEBUG oslo_concurrency.lockutils [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Acquired lock "refresh_cache-59d2ee3a-756f-4dda-a70a-6d43aaf32049" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.472790] env[61867]: DEBUG nova.network.neutron [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 689.788737] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquiring lock "37ed9e18-8dba-459d-bc67-e3ce3f9cdaea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.788737] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Lock "37ed9e18-8dba-459d-bc67-e3ce3f9cdaea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.788737] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquiring lock "37ed9e18-8dba-459d-bc67-e3ce3f9cdaea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.788854] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Lock "37ed9e18-8dba-459d-bc67-e3ce3f9cdaea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.790020] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Lock "37ed9e18-8dba-459d-bc67-e3ce3f9cdaea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.794929] env[61867]: INFO nova.compute.manager [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Terminating instance [ 689.796504] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquiring lock "refresh_cache-37ed9e18-8dba-459d-bc67-e3ce3f9cdaea" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.796718] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquired lock "refresh_cache-37ed9e18-8dba-459d-bc67-e3ce3f9cdaea" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.796904] env[61867]: DEBUG nova.network.neutron [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 689.846663] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "refresh_cache-274c5625-eb81-45d1-bd95-7336bddfad1e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.846663] env[61867]: DEBUG nova.compute.manager [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 689.846663] env[61867]: DEBUG nova.compute.manager [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 689.846663] env[61867]: DEBUG nova.network.neutron [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 689.867206] env[61867]: DEBUG nova.network.neutron [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.994100] env[61867]: DEBUG nova.network.neutron [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.074344] env[61867]: DEBUG nova.network.neutron [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.091611] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f353da67-8289-4e35-ae76-ce152834e0bd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.100709] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52cc16ef-33b2-441d-8973-6ff3294a200c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.132950] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49202998-2ac8-48d3-ad45-2192c6d51657 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.145021] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee6df241-36ae-42e7-844a-c4b56dd852db {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.158474] env[61867]: DEBUG nova.compute.provider_tree [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.327554] env[61867]: DEBUG nova.network.neutron [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.372144] env[61867]: DEBUG nova.network.neutron [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.383553] env[61867]: DEBUG nova.compute.manager [req-bb9f7191-16e7-4bda-861f-a960f9515e56 req-f23f29a1-99fb-41f4-9e60-e3ecde090c97 service nova] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Received event network-vif-deleted-78d19f74-61d7-47cc-bafd-33d2683cb241 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 690.395985] env[61867]: DEBUG nova.network.neutron [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.583024] env[61867]: DEBUG oslo_concurrency.lockutils [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Releasing lock "refresh_cache-59d2ee3a-756f-4dda-a70a-6d43aaf32049" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.583024] env[61867]: DEBUG nova.compute.manager [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 690.583024] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 690.583024] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8a94580f-a8ed-4312-a6d0-28639eaf138d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.591040] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d9f139-d2f1-48b4-a8fe-34799d9c0c99 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.619529] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 59d2ee3a-756f-4dda-a70a-6d43aaf32049 could not be found. [ 690.619920] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 690.620190] env[61867]: INFO nova.compute.manager [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Took 0.04 seconds to destroy the instance on the hypervisor. [ 690.620621] env[61867]: DEBUG oslo.service.loopingcall [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 690.620945] env[61867]: DEBUG nova.compute.manager [-] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 690.621056] env[61867]: DEBUG nova.network.neutron [-] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 690.637619] env[61867]: DEBUG nova.network.neutron [-] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.661861] env[61867]: DEBUG nova.scheduler.client.report [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 690.875396] env[61867]: INFO nova.compute.manager [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 274c5625-eb81-45d1-bd95-7336bddfad1e] Took 1.03 seconds to deallocate network for instance. [ 690.905200] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Releasing lock "refresh_cache-37ed9e18-8dba-459d-bc67-e3ce3f9cdaea" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.905200] env[61867]: DEBUG nova.compute.manager [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 690.905200] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 690.905200] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b31f26be-e72e-470d-9254-0f7c9149ea2d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.915835] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 690.916895] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7a93122-eecc-41a0-abcc-d166a7f0a480 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.927782] env[61867]: DEBUG oslo_vmware.api [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 690.927782] env[61867]: value = "task-1276362" [ 690.927782] env[61867]: _type = "Task" [ 690.927782] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.943563] env[61867]: DEBUG oslo_vmware.api [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276362, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.140948] env[61867]: DEBUG nova.network.neutron [-] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.167232] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.490s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.170923] env[61867]: DEBUG nova.compute.manager [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 691.178136] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.287s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.439518] env[61867]: DEBUG oslo_vmware.api [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276362, 'name': PowerOffVM_Task, 'duration_secs': 0.118612} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.442143] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 691.442411] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 691.442752] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d356f200-cde6-46f0-9d2b-b5368987eb8f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.465686] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 691.465915] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 691.466105] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Deleting the datastore file [datastore2] 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 691.466360] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-309a224f-620d-42c7-8344-a8ca3d3abce3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.472765] env[61867]: DEBUG oslo_vmware.api [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for the task: (returnval){ [ 691.472765] env[61867]: value = "task-1276364" [ 691.472765] env[61867]: _type = "Task" [ 691.472765] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.482197] env[61867]: DEBUG oslo_vmware.api [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276364, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.642586] env[61867]: INFO nova.compute.manager [-] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Took 1.02 seconds to deallocate network for instance. [ 691.645183] env[61867]: DEBUG nova.compute.claims [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 691.645259] env[61867]: DEBUG oslo_concurrency.lockutils [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.687355] env[61867]: DEBUG nova.compute.utils [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 691.698310] env[61867]: DEBUG nova.compute.manager [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 691.698310] env[61867]: DEBUG nova.network.neutron [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 691.768716] env[61867]: DEBUG nova.policy [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b56955becec849dca65ab396c3c67402', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2e8433d5373245dea17aa7c98fbb0a58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 691.921179] env[61867]: INFO nova.scheduler.client.report [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleted allocations for instance 274c5625-eb81-45d1-bd95-7336bddfad1e [ 691.984191] env[61867]: DEBUG oslo_vmware.api [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Task: {'id': task-1276364, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.088077} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.984464] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 691.984660] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 691.985549] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 691.985549] env[61867]: INFO nova.compute.manager [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Took 1.08 seconds to destroy the instance on the hypervisor. [ 691.985549] env[61867]: DEBUG oslo.service.loopingcall [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 691.985549] env[61867]: DEBUG nova.compute.manager [-] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 691.985778] env[61867]: DEBUG nova.network.neutron [-] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 692.026727] env[61867]: DEBUG nova.network.neutron [-] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.151779] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69df0aea-7416-46b2-b1fb-7ae15c1bc180 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.160325] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27ee2f75-6caf-4b12-94ca-0f5e85d16597 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.189353] env[61867]: DEBUG nova.network.neutron [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Successfully created port: 19b41515-c7df-4409-9145-b398f19f6eec {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 692.192373] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b2515c-6761-424e-917b-0062266af841 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.199971] env[61867]: DEBUG nova.compute.manager [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 692.203738] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a86641c-f34d-4cf3-ab82-a742aba0f694 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.218754] env[61867]: DEBUG nova.compute.provider_tree [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.433372] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a078b947-1c43-4c9a-91e1-530326ec958d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "274c5625-eb81-45d1-bd95-7336bddfad1e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.333s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.530616] env[61867]: DEBUG nova.network.neutron [-] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.723987] env[61867]: DEBUG nova.scheduler.client.report [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.938289] env[61867]: DEBUG nova.compute.manager [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 693.036571] env[61867]: INFO nova.compute.manager [-] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Took 1.05 seconds to deallocate network for instance. [ 693.098128] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "3022ea79-7c27-40a4-80d6-1a86b354917a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.098794] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "3022ea79-7c27-40a4-80d6-1a86b354917a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.160188] env[61867]: DEBUG nova.compute.manager [req-531ca3a2-136f-496d-aed8-0272f42a45ba req-dd28691e-15ab-4a8f-8037-bb9b32cba20b service nova] [instance: 774e7c38-426a-441d-8252-3daa424be558] Received event network-changed-19b41515-c7df-4409-9145-b398f19f6eec {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 693.160188] env[61867]: DEBUG nova.compute.manager [req-531ca3a2-136f-496d-aed8-0272f42a45ba req-dd28691e-15ab-4a8f-8037-bb9b32cba20b service nova] [instance: 774e7c38-426a-441d-8252-3daa424be558] Refreshing instance network info cache due to event network-changed-19b41515-c7df-4409-9145-b398f19f6eec. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 693.160188] env[61867]: DEBUG oslo_concurrency.lockutils [req-531ca3a2-136f-496d-aed8-0272f42a45ba req-dd28691e-15ab-4a8f-8037-bb9b32cba20b service nova] Acquiring lock "refresh_cache-774e7c38-426a-441d-8252-3daa424be558" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.160188] env[61867]: DEBUG oslo_concurrency.lockutils [req-531ca3a2-136f-496d-aed8-0272f42a45ba req-dd28691e-15ab-4a8f-8037-bb9b32cba20b service nova] Acquired lock "refresh_cache-774e7c38-426a-441d-8252-3daa424be558" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.160188] env[61867]: DEBUG nova.network.neutron [req-531ca3a2-136f-496d-aed8-0272f42a45ba req-dd28691e-15ab-4a8f-8037-bb9b32cba20b service nova] [instance: 774e7c38-426a-441d-8252-3daa424be558] Refreshing network info cache for port 19b41515-c7df-4409-9145-b398f19f6eec {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 693.173425] env[61867]: ERROR nova.compute.manager [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 19b41515-c7df-4409-9145-b398f19f6eec, please check neutron logs for more information. [ 693.173425] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 693.173425] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.173425] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 693.173425] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.173425] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 693.173425] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.173425] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 693.173425] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.173425] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 693.173425] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.173425] env[61867]: ERROR nova.compute.manager raise self.value [ 693.173425] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.173425] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 693.173425] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.173425] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 693.173927] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.173927] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 693.173927] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 19b41515-c7df-4409-9145-b398f19f6eec, please check neutron logs for more information. [ 693.173927] env[61867]: ERROR nova.compute.manager [ 693.173927] env[61867]: Traceback (most recent call last): [ 693.173927] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 693.173927] env[61867]: listener.cb(fileno) [ 693.173927] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.173927] env[61867]: result = function(*args, **kwargs) [ 693.173927] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 693.173927] env[61867]: return func(*args, **kwargs) [ 693.173927] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.173927] env[61867]: raise e [ 693.173927] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.173927] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 693.173927] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.173927] env[61867]: created_port_ids = self._update_ports_for_instance( [ 693.173927] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.173927] env[61867]: with excutils.save_and_reraise_exception(): [ 693.173927] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.173927] env[61867]: self.force_reraise() [ 693.173927] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.173927] env[61867]: raise self.value [ 693.173927] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.173927] env[61867]: updated_port = self._update_port( [ 693.173927] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.173927] env[61867]: _ensure_no_port_binding_failure(port) [ 693.173927] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.173927] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 693.174759] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 19b41515-c7df-4409-9145-b398f19f6eec, please check neutron logs for more information. [ 693.174759] env[61867]: Removing descriptor: 21 [ 693.216736] env[61867]: DEBUG nova.compute.manager [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 693.233664] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.061s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.233664] env[61867]: ERROR nova.compute.manager [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7ab0af92-a6e0-46c0-bbf5-bda124b94276, please check neutron logs for more information. [ 693.233664] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Traceback (most recent call last): [ 693.233664] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 693.233664] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] self.driver.spawn(context, instance, image_meta, [ 693.233664] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 693.233664] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 693.233664] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 693.233664] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] vm_ref = self.build_virtual_machine(instance, [ 693.234078] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 693.234078] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 693.234078] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 693.234078] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] for vif in network_info: [ 693.234078] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 693.234078] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] return self._sync_wrapper(fn, *args, **kwargs) [ 693.234078] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 693.234078] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] self.wait() [ 693.234078] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 693.234078] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] self[:] = self._gt.wait() [ 693.234078] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 693.234078] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] return self._exit_event.wait() [ 693.234078] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 693.234501] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] current.throw(*self._exc) [ 693.234501] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.234501] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] result = function(*args, **kwargs) [ 693.234501] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 693.234501] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] return func(*args, **kwargs) [ 693.234501] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.234501] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] raise e [ 693.234501] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.234501] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] nwinfo = self.network_api.allocate_for_instance( [ 693.234501] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.234501] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] created_port_ids = self._update_ports_for_instance( [ 693.234501] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.234501] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] with excutils.save_and_reraise_exception(): [ 693.234886] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.234886] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] self.force_reraise() [ 693.234886] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.234886] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] raise self.value [ 693.234886] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.234886] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] updated_port = self._update_port( [ 693.234886] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.234886] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] _ensure_no_port_binding_failure(port) [ 693.234886] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.234886] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] raise exception.PortBindingFailed(port_id=port['id']) [ 693.234886] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] nova.exception.PortBindingFailed: Binding failed for port 7ab0af92-a6e0-46c0-bbf5-bda124b94276, please check neutron logs for more information. [ 693.234886] env[61867]: ERROR nova.compute.manager [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] [ 693.235218] env[61867]: DEBUG nova.compute.utils [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Binding failed for port 7ab0af92-a6e0-46c0-bbf5-bda124b94276, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 693.236059] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.123s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.237627] env[61867]: INFO nova.compute.claims [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 693.245044] env[61867]: DEBUG nova.compute.manager [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Build of instance aad1e5ea-5a9c-4499-8f67-2516c5c285cc was re-scheduled: Binding failed for port 7ab0af92-a6e0-46c0-bbf5-bda124b94276, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 693.245675] env[61867]: DEBUG nova.compute.manager [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 693.245771] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "refresh_cache-aad1e5ea-5a9c-4499-8f67-2516c5c285cc" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.245859] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquired lock "refresh_cache-aad1e5ea-5a9c-4499-8f67-2516c5c285cc" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.246030] env[61867]: DEBUG nova.network.neutron [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 693.253106] env[61867]: DEBUG nova.virt.hardware [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 693.253292] env[61867]: DEBUG nova.virt.hardware [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 693.253450] env[61867]: DEBUG nova.virt.hardware [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 693.253629] env[61867]: DEBUG nova.virt.hardware [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 693.253770] env[61867]: DEBUG nova.virt.hardware [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 693.253908] env[61867]: DEBUG nova.virt.hardware [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 693.254158] env[61867]: DEBUG nova.virt.hardware [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 693.254332] env[61867]: DEBUG nova.virt.hardware [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 693.254479] env[61867]: DEBUG nova.virt.hardware [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 693.254729] env[61867]: DEBUG nova.virt.hardware [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 693.254805] env[61867]: DEBUG nova.virt.hardware [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 693.257656] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-393090db-618a-4aa6-a82a-000427b1d319 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.264388] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-033e386c-0004-43f1-b799-22f07a02d2c9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.288127] env[61867]: ERROR nova.compute.manager [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 19b41515-c7df-4409-9145-b398f19f6eec, please check neutron logs for more information. [ 693.288127] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] Traceback (most recent call last): [ 693.288127] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 693.288127] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] yield resources [ 693.288127] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 693.288127] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] self.driver.spawn(context, instance, image_meta, [ 693.288127] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 693.288127] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] self._vmops.spawn(context, instance, image_meta, injected_files, [ 693.288127] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 693.288127] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] vm_ref = self.build_virtual_machine(instance, [ 693.288127] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 693.288526] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] vif_infos = vmwarevif.get_vif_info(self._session, [ 693.288526] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 693.288526] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] for vif in network_info: [ 693.288526] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 693.288526] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] return self._sync_wrapper(fn, *args, **kwargs) [ 693.288526] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 693.288526] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] self.wait() [ 693.288526] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 693.288526] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] self[:] = self._gt.wait() [ 693.288526] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 693.288526] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] return self._exit_event.wait() [ 693.288526] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 693.288526] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] current.throw(*self._exc) [ 693.288883] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.288883] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] result = function(*args, **kwargs) [ 693.288883] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 693.288883] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] return func(*args, **kwargs) [ 693.288883] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.288883] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] raise e [ 693.288883] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.288883] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] nwinfo = self.network_api.allocate_for_instance( [ 693.288883] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.288883] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] created_port_ids = self._update_ports_for_instance( [ 693.288883] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.288883] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] with excutils.save_and_reraise_exception(): [ 693.288883] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.289238] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] self.force_reraise() [ 693.289238] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.289238] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] raise self.value [ 693.289238] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.289238] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] updated_port = self._update_port( [ 693.289238] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.289238] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] _ensure_no_port_binding_failure(port) [ 693.289238] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.289238] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] raise exception.PortBindingFailed(port_id=port['id']) [ 693.289238] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] nova.exception.PortBindingFailed: Binding failed for port 19b41515-c7df-4409-9145-b398f19f6eec, please check neutron logs for more information. [ 693.289238] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] [ 693.289238] env[61867]: INFO nova.compute.manager [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Terminating instance [ 693.293020] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Acquiring lock "refresh_cache-774e7c38-426a-441d-8252-3daa424be558" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.458110] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.546356] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.682828] env[61867]: DEBUG nova.network.neutron [req-531ca3a2-136f-496d-aed8-0272f42a45ba req-dd28691e-15ab-4a8f-8037-bb9b32cba20b service nova] [instance: 774e7c38-426a-441d-8252-3daa424be558] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.763704] env[61867]: DEBUG nova.network.neutron [req-531ca3a2-136f-496d-aed8-0272f42a45ba req-dd28691e-15ab-4a8f-8037-bb9b32cba20b service nova] [instance: 774e7c38-426a-441d-8252-3daa424be558] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.778404] env[61867]: DEBUG nova.network.neutron [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.860716] env[61867]: DEBUG nova.network.neutron [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.267084] env[61867]: DEBUG oslo_concurrency.lockutils [req-531ca3a2-136f-496d-aed8-0272f42a45ba req-dd28691e-15ab-4a8f-8037-bb9b32cba20b service nova] Releasing lock "refresh_cache-774e7c38-426a-441d-8252-3daa424be558" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.267508] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Acquired lock "refresh_cache-774e7c38-426a-441d-8252-3daa424be558" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.267732] env[61867]: DEBUG nova.network.neutron [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 694.363274] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Releasing lock "refresh_cache-aad1e5ea-5a9c-4499-8f67-2516c5c285cc" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.363517] env[61867]: DEBUG nova.compute.manager [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 694.363584] env[61867]: DEBUG nova.compute.manager [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 694.363728] env[61867]: DEBUG nova.network.neutron [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 694.379956] env[61867]: DEBUG nova.network.neutron [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.539919] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9acc0030-c59c-4eee-b2c5-ea4c5dd2e76b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.548140] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef35f6d-3cd1-42c9-9d92-74ae6217c4c6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.577382] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-498fde55-5cfa-432c-bbb8-7c3318ba64e3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.584462] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672f9149-97da-473b-981f-985e42d5461a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.597547] env[61867]: DEBUG nova.compute.provider_tree [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.760027] env[61867]: DEBUG oslo_concurrency.lockutils [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "be825f62-cb56-4647-be1c-1605536fdc89" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.760131] env[61867]: DEBUG oslo_concurrency.lockutils [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "be825f62-cb56-4647-be1c-1605536fdc89" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.785117] env[61867]: DEBUG nova.network.neutron [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.854925] env[61867]: DEBUG nova.network.neutron [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.882807] env[61867]: DEBUG nova.network.neutron [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.100644] env[61867]: DEBUG nova.scheduler.client.report [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 695.183611] env[61867]: DEBUG nova.compute.manager [req-223a3cda-9663-4a4c-9e1a-dfaf6a98566e req-d4b1516f-868d-4463-b932-7159f2e3db5c service nova] [instance: 774e7c38-426a-441d-8252-3daa424be558] Received event network-vif-deleted-19b41515-c7df-4409-9145-b398f19f6eec {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 695.357875] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Releasing lock "refresh_cache-774e7c38-426a-441d-8252-3daa424be558" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.358307] env[61867]: DEBUG nova.compute.manager [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 695.358508] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 695.358875] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ce22cd21-dd19-4af9-925c-70d87f2d1b03 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.368125] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d5b9d8-04ba-4fab-94c4-af51b8dea892 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.388522] env[61867]: INFO nova.compute.manager [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: aad1e5ea-5a9c-4499-8f67-2516c5c285cc] Took 1.02 seconds to deallocate network for instance. [ 695.391241] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 774e7c38-426a-441d-8252-3daa424be558 could not be found. [ 695.391343] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 695.391513] env[61867]: INFO nova.compute.manager [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Took 0.03 seconds to destroy the instance on the hypervisor. [ 695.391746] env[61867]: DEBUG oslo.service.loopingcall [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 695.392145] env[61867]: DEBUG nova.compute.manager [-] [instance: 774e7c38-426a-441d-8252-3daa424be558] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 695.392240] env[61867]: DEBUG nova.network.neutron [-] [instance: 774e7c38-426a-441d-8252-3daa424be558] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 695.407092] env[61867]: DEBUG nova.network.neutron [-] [instance: 774e7c38-426a-441d-8252-3daa424be558] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.606028] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.606028] env[61867]: DEBUG nova.compute.manager [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 695.608512] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.755s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.609956] env[61867]: INFO nova.compute.claims [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 695.909166] env[61867]: DEBUG nova.network.neutron [-] [instance: 774e7c38-426a-441d-8252-3daa424be558] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.114447] env[61867]: DEBUG nova.compute.utils [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 696.118097] env[61867]: DEBUG nova.compute.manager [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 696.118269] env[61867]: DEBUG nova.network.neutron [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 696.164311] env[61867]: DEBUG nova.policy [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ef44a9804b62462aa2d360311c0d860e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8794e43cea1b497c8e5559045ac693ce', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 696.411363] env[61867]: INFO nova.compute.manager [-] [instance: 774e7c38-426a-441d-8252-3daa424be558] Took 1.02 seconds to deallocate network for instance. [ 696.415398] env[61867]: DEBUG nova.compute.claims [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 696.415398] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.425889] env[61867]: INFO nova.scheduler.client.report [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Deleted allocations for instance aad1e5ea-5a9c-4499-8f67-2516c5c285cc [ 696.456026] env[61867]: DEBUG nova.network.neutron [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Successfully created port: f0f2c586-6ec6-41cf-8afc-6e19ae0ae039 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 696.619148] env[61867]: DEBUG nova.compute.manager [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 696.920041] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbf42535-6df1-444c-a26d-ac71a9576ee0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.928200] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4056bdbd-6626-4b1e-b390-8d4f76271e3f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.958200] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20bc07be-d97e-4ec5-b9b4-0812ffea920e tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "aad1e5ea-5a9c-4499-8f67-2516c5c285cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.539s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.960755] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8603eb-9290-4763-9233-374669e7583f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.979881] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c34a78-6601-4f0c-80f2-b53652d84b7f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.996251] env[61867]: DEBUG nova.compute.provider_tree [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 697.127019] env[61867]: INFO nova.virt.block_device [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Booting with volume bf7183e2-a24a-4ce6-9ea1-7909e196ef19 at /dev/sda [ 697.164729] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bee92123-84b5-42a0-adcb-c65a25da39c2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.173290] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8cc48f8-1b96-4ed6-8275-23e75be3553a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.193764] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7f1fde7a-d18e-48ba-b7b3-fe613e3d0d84 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.201376] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdcccdea-247a-42d1-bd40-fa061c4fd319 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.220110] env[61867]: DEBUG nova.compute.manager [req-fba4d66f-e7d7-4cfa-a994-26e64ae93c8a req-be99c6e4-e687-468c-b01a-9feffabeb191 service nova] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Received event network-changed-f0f2c586-6ec6-41cf-8afc-6e19ae0ae039 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 697.220314] env[61867]: DEBUG nova.compute.manager [req-fba4d66f-e7d7-4cfa-a994-26e64ae93c8a req-be99c6e4-e687-468c-b01a-9feffabeb191 service nova] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Refreshing instance network info cache due to event network-changed-f0f2c586-6ec6-41cf-8afc-6e19ae0ae039. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 697.220522] env[61867]: DEBUG oslo_concurrency.lockutils [req-fba4d66f-e7d7-4cfa-a994-26e64ae93c8a req-be99c6e4-e687-468c-b01a-9feffabeb191 service nova] Acquiring lock "refresh_cache-69d292d9-2330-47ac-94d4-6797abb8c167" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.220681] env[61867]: DEBUG oslo_concurrency.lockutils [req-fba4d66f-e7d7-4cfa-a994-26e64ae93c8a req-be99c6e4-e687-468c-b01a-9feffabeb191 service nova] Acquired lock "refresh_cache-69d292d9-2330-47ac-94d4-6797abb8c167" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.220808] env[61867]: DEBUG nova.network.neutron [req-fba4d66f-e7d7-4cfa-a994-26e64ae93c8a req-be99c6e4-e687-468c-b01a-9feffabeb191 service nova] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Refreshing network info cache for port f0f2c586-6ec6-41cf-8afc-6e19ae0ae039 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 697.226161] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63fe44f3-a25e-438d-a61c-8b1c2540df56 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.233624] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb0c0a94-5112-45cf-8321-ce83614bec91 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.247459] env[61867]: DEBUG nova.virt.block_device [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Updating existing volume attachment record: 5b32140d-e426-4b95-9e77-d6825e6cf408 {{(pid=61867) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 697.312224] env[61867]: ERROR nova.compute.manager [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f0f2c586-6ec6-41cf-8afc-6e19ae0ae039, please check neutron logs for more information. [ 697.312224] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 697.312224] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.312224] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 697.312224] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 697.312224] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 697.312224] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 697.312224] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 697.312224] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.312224] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 697.312224] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.312224] env[61867]: ERROR nova.compute.manager raise self.value [ 697.312224] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 697.312224] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 697.312224] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.312224] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 697.312926] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.312926] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 697.312926] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f0f2c586-6ec6-41cf-8afc-6e19ae0ae039, please check neutron logs for more information. [ 697.312926] env[61867]: ERROR nova.compute.manager [ 697.312926] env[61867]: Traceback (most recent call last): [ 697.312926] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 697.312926] env[61867]: listener.cb(fileno) [ 697.312926] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.312926] env[61867]: result = function(*args, **kwargs) [ 697.312926] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 697.312926] env[61867]: return func(*args, **kwargs) [ 697.312926] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.312926] env[61867]: raise e [ 697.312926] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.312926] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 697.312926] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 697.312926] env[61867]: created_port_ids = self._update_ports_for_instance( [ 697.312926] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 697.312926] env[61867]: with excutils.save_and_reraise_exception(): [ 697.312926] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.312926] env[61867]: self.force_reraise() [ 697.312926] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.312926] env[61867]: raise self.value [ 697.312926] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 697.312926] env[61867]: updated_port = self._update_port( [ 697.312926] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.312926] env[61867]: _ensure_no_port_binding_failure(port) [ 697.312926] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.312926] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 697.314082] env[61867]: nova.exception.PortBindingFailed: Binding failed for port f0f2c586-6ec6-41cf-8afc-6e19ae0ae039, please check neutron logs for more information. [ 697.314082] env[61867]: Removing descriptor: 17 [ 697.474383] env[61867]: DEBUG nova.compute.manager [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 697.502378] env[61867]: DEBUG nova.scheduler.client.report [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 697.742695] env[61867]: DEBUG nova.network.neutron [req-fba4d66f-e7d7-4cfa-a994-26e64ae93c8a req-be99c6e4-e687-468c-b01a-9feffabeb191 service nova] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.821339] env[61867]: DEBUG nova.network.neutron [req-fba4d66f-e7d7-4cfa-a994-26e64ae93c8a req-be99c6e4-e687-468c-b01a-9feffabeb191 service nova] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.993171] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.008083] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.400s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.008590] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 698.011131] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.159s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.324793] env[61867]: DEBUG oslo_concurrency.lockutils [req-fba4d66f-e7d7-4cfa-a994-26e64ae93c8a req-be99c6e4-e687-468c-b01a-9feffabeb191 service nova] Releasing lock "refresh_cache-69d292d9-2330-47ac-94d4-6797abb8c167" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.344947] env[61867]: DEBUG oslo_concurrency.lockutils [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "80a703b3-d692-4023-a73b-dba980a94dff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.345212] env[61867]: DEBUG oslo_concurrency.lockutils [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "80a703b3-d692-4023-a73b-dba980a94dff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.515557] env[61867]: DEBUG nova.compute.utils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 698.518172] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 698.518172] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 698.567946] env[61867]: DEBUG nova.policy [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d7b6b0de3374b09afb54bbc5c7c2382', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a47120d2cb044db0bb45a9e801d1fe2d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 698.823992] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f5192b-3052-4bcb-8d69-88fc589695d9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.830745] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Successfully created port: 0ccf49d0-6d5d-491b-a98a-d0afda9b598c {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 698.836307] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64136f26-02ec-47d0-b4fa-ee68040c50b1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.871142] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89295bba-bae5-4fd4-b040-195d43d4b6c6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.878643] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e2809a-7183-45de-8df2-f21de819004b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.894143] env[61867]: DEBUG nova.compute.provider_tree [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 699.018675] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 699.244711] env[61867]: DEBUG nova.compute.manager [req-f871fef6-8a1b-446c-ae44-178737bffce0 req-3e748d07-20da-4617-ad31-737b3c28829e service nova] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Received event network-vif-deleted-f0f2c586-6ec6-41cf-8afc-6e19ae0ae039 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 699.372217] env[61867]: DEBUG nova.compute.manager [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 699.372758] env[61867]: DEBUG nova.virt.hardware [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 699.373038] env[61867]: DEBUG nova.virt.hardware [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 699.373255] env[61867]: DEBUG nova.virt.hardware [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 699.373454] env[61867]: DEBUG nova.virt.hardware [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 699.373597] env[61867]: DEBUG nova.virt.hardware [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 699.373740] env[61867]: DEBUG nova.virt.hardware [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 699.373938] env[61867]: DEBUG nova.virt.hardware [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 699.374100] env[61867]: DEBUG nova.virt.hardware [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 699.375043] env[61867]: DEBUG nova.virt.hardware [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 699.375043] env[61867]: DEBUG nova.virt.hardware [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 699.375043] env[61867]: DEBUG nova.virt.hardware [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.375459] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eabf1ed-2364-4bea-a5d6-87d781a38fa6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.384410] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca7a8ed-16cf-4e5d-9717-6f44ffdad94e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.398857] env[61867]: DEBUG nova.scheduler.client.report [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.402572] env[61867]: ERROR nova.compute.manager [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f0f2c586-6ec6-41cf-8afc-6e19ae0ae039, please check neutron logs for more information. [ 699.402572] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Traceback (most recent call last): [ 699.402572] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 699.402572] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] yield resources [ 699.402572] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.402572] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] self.driver.spawn(context, instance, image_meta, [ 699.402572] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 699.402572] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.402572] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.402572] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] vm_ref = self.build_virtual_machine(instance, [ 699.402572] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.402950] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.402950] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.402950] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] for vif in network_info: [ 699.402950] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.402950] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] return self._sync_wrapper(fn, *args, **kwargs) [ 699.402950] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.402950] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] self.wait() [ 699.402950] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.402950] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] self[:] = self._gt.wait() [ 699.402950] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.402950] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] return self._exit_event.wait() [ 699.402950] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 699.402950] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] current.throw(*self._exc) [ 699.403365] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.403365] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] result = function(*args, **kwargs) [ 699.403365] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 699.403365] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] return func(*args, **kwargs) [ 699.403365] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.403365] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] raise e [ 699.403365] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.403365] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] nwinfo = self.network_api.allocate_for_instance( [ 699.403365] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.403365] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] created_port_ids = self._update_ports_for_instance( [ 699.403365] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.403365] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] with excutils.save_and_reraise_exception(): [ 699.403365] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.403692] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] self.force_reraise() [ 699.403692] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.403692] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] raise self.value [ 699.403692] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.403692] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] updated_port = self._update_port( [ 699.403692] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.403692] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] _ensure_no_port_binding_failure(port) [ 699.403692] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.403692] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] raise exception.PortBindingFailed(port_id=port['id']) [ 699.403692] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] nova.exception.PortBindingFailed: Binding failed for port f0f2c586-6ec6-41cf-8afc-6e19ae0ae039, please check neutron logs for more information. [ 699.403692] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] [ 699.403692] env[61867]: INFO nova.compute.manager [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Terminating instance [ 699.405708] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Acquiring lock "refresh_cache-69d292d9-2330-47ac-94d4-6797abb8c167" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.405863] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Acquired lock "refresh_cache-69d292d9-2330-47ac-94d4-6797abb8c167" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.406040] env[61867]: DEBUG nova.network.neutron [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 699.658523] env[61867]: ERROR nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0ccf49d0-6d5d-491b-a98a-d0afda9b598c, please check neutron logs for more information. [ 699.658523] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 699.658523] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.658523] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 699.658523] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.658523] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 699.658523] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.658523] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 699.658523] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.658523] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 699.658523] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.658523] env[61867]: ERROR nova.compute.manager raise self.value [ 699.658523] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.658523] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 699.658523] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.658523] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 699.659295] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.659295] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 699.659295] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0ccf49d0-6d5d-491b-a98a-d0afda9b598c, please check neutron logs for more information. [ 699.659295] env[61867]: ERROR nova.compute.manager [ 699.659295] env[61867]: Traceback (most recent call last): [ 699.659295] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 699.659295] env[61867]: listener.cb(fileno) [ 699.659295] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.659295] env[61867]: result = function(*args, **kwargs) [ 699.659295] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 699.659295] env[61867]: return func(*args, **kwargs) [ 699.659295] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.659295] env[61867]: raise e [ 699.659295] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.659295] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 699.659295] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.659295] env[61867]: created_port_ids = self._update_ports_for_instance( [ 699.659295] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.659295] env[61867]: with excutils.save_and_reraise_exception(): [ 699.659295] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.659295] env[61867]: self.force_reraise() [ 699.659295] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.659295] env[61867]: raise self.value [ 699.659295] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.659295] env[61867]: updated_port = self._update_port( [ 699.659295] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.659295] env[61867]: _ensure_no_port_binding_failure(port) [ 699.659295] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.659295] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 699.660091] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 0ccf49d0-6d5d-491b-a98a-d0afda9b598c, please check neutron logs for more information. [ 699.660091] env[61867]: Removing descriptor: 17 [ 699.906586] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.895s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.907248] env[61867]: ERROR nova.compute.manager [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5ec8abc2-e15e-4ef5-94d2-3a2c6d0740e4, please check neutron logs for more information. [ 699.907248] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] Traceback (most recent call last): [ 699.907248] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.907248] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] self.driver.spawn(context, instance, image_meta, [ 699.907248] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 699.907248] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.907248] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.907248] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] vm_ref = self.build_virtual_machine(instance, [ 699.907248] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.907248] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.907248] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.907578] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] for vif in network_info: [ 699.907578] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.907578] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] return self._sync_wrapper(fn, *args, **kwargs) [ 699.907578] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.907578] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] self.wait() [ 699.907578] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.907578] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] self[:] = self._gt.wait() [ 699.907578] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.907578] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] return self._exit_event.wait() [ 699.907578] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 699.907578] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] current.throw(*self._exc) [ 699.907578] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.907578] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] result = function(*args, **kwargs) [ 699.907904] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 699.907904] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] return func(*args, **kwargs) [ 699.907904] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.907904] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] raise e [ 699.907904] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.907904] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] nwinfo = self.network_api.allocate_for_instance( [ 699.907904] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.907904] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] created_port_ids = self._update_ports_for_instance( [ 699.907904] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.907904] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] with excutils.save_and_reraise_exception(): [ 699.907904] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.907904] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] self.force_reraise() [ 699.907904] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.908240] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] raise self.value [ 699.908240] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.908240] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] updated_port = self._update_port( [ 699.908240] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.908240] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] _ensure_no_port_binding_failure(port) [ 699.908240] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.908240] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] raise exception.PortBindingFailed(port_id=port['id']) [ 699.908240] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] nova.exception.PortBindingFailed: Binding failed for port 5ec8abc2-e15e-4ef5-94d2-3a2c6d0740e4, please check neutron logs for more information. [ 699.908240] env[61867]: ERROR nova.compute.manager [instance: 5c250675-3708-4c25-b370-6707c457357a] [ 699.908240] env[61867]: DEBUG nova.compute.utils [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Binding failed for port 5ec8abc2-e15e-4ef5-94d2-3a2c6d0740e4, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 699.909591] env[61867]: DEBUG oslo_concurrency.lockutils [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.347s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.915720] env[61867]: DEBUG nova.compute.manager [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Build of instance 5c250675-3708-4c25-b370-6707c457357a was re-scheduled: Binding failed for port 5ec8abc2-e15e-4ef5-94d2-3a2c6d0740e4, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 699.915720] env[61867]: DEBUG nova.compute.manager [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 699.915720] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "refresh_cache-5c250675-3708-4c25-b370-6707c457357a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.915720] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquired lock "refresh_cache-5c250675-3708-4c25-b370-6707c457357a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.915935] env[61867]: DEBUG nova.network.neutron [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 699.928882] env[61867]: DEBUG nova.network.neutron [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.002691] env[61867]: DEBUG nova.network.neutron [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.033255] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 700.056660] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 700.056906] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 700.057078] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 700.057259] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 700.057403] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 700.057548] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 700.057752] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 700.057910] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 700.058085] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 700.058246] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 700.058414] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 700.059301] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf9773d-b404-4e3e-97f7-4b543dd9030b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.066994] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bdb737b-e87a-4c12-a2f0-75f853203a39 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.080367] env[61867]: ERROR nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0ccf49d0-6d5d-491b-a98a-d0afda9b598c, please check neutron logs for more information. [ 700.080367] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Traceback (most recent call last): [ 700.080367] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 700.080367] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] yield resources [ 700.080367] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 700.080367] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] self.driver.spawn(context, instance, image_meta, [ 700.080367] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 700.080367] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 700.080367] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 700.080367] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] vm_ref = self.build_virtual_machine(instance, [ 700.080367] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 700.080768] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 700.080768] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 700.080768] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] for vif in network_info: [ 700.080768] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 700.080768] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] return self._sync_wrapper(fn, *args, **kwargs) [ 700.080768] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 700.080768] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] self.wait() [ 700.080768] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 700.080768] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] self[:] = self._gt.wait() [ 700.080768] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 700.080768] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] return self._exit_event.wait() [ 700.080768] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 700.080768] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] current.throw(*self._exc) [ 700.081174] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 700.081174] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] result = function(*args, **kwargs) [ 700.081174] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 700.081174] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] return func(*args, **kwargs) [ 700.081174] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 700.081174] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] raise e [ 700.081174] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 700.081174] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] nwinfo = self.network_api.allocate_for_instance( [ 700.081174] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 700.081174] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] created_port_ids = self._update_ports_for_instance( [ 700.081174] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 700.081174] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] with excutils.save_and_reraise_exception(): [ 700.081174] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 700.081566] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] self.force_reraise() [ 700.081566] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 700.081566] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] raise self.value [ 700.081566] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 700.081566] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] updated_port = self._update_port( [ 700.081566] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 700.081566] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] _ensure_no_port_binding_failure(port) [ 700.081566] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 700.081566] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] raise exception.PortBindingFailed(port_id=port['id']) [ 700.081566] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] nova.exception.PortBindingFailed: Binding failed for port 0ccf49d0-6d5d-491b-a98a-d0afda9b598c, please check neutron logs for more information. [ 700.081566] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] [ 700.081566] env[61867]: INFO nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Terminating instance [ 700.082724] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquiring lock "refresh_cache-3f1696b0-4de4-4128-bc83-e539e48dc8e3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.082880] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquired lock "refresh_cache-3f1696b0-4de4-4128-bc83-e539e48dc8e3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.083053] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 700.434601] env[61867]: DEBUG nova.network.neutron [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.505783] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Releasing lock "refresh_cache-69d292d9-2330-47ac-94d4-6797abb8c167" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.506283] env[61867]: DEBUG nova.compute.manager [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 700.506596] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-23c9d05d-204b-401f-86f2-cd4f4edca0b0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.510960] env[61867]: DEBUG nova.network.neutron [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.516057] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e2177cc-d8ef-4b85-97a6-bb2810869930 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.542261] env[61867]: WARNING nova.virt.vmwareapi.driver [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 69d292d9-2330-47ac-94d4-6797abb8c167 could not be found. [ 700.543065] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 700.545159] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7994b99c-cf39-468a-b818-e35efeeb08ba {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.552864] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9759565c-1855-4ff2-9be6-e58a4d4a46cb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.575685] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 69d292d9-2330-47ac-94d4-6797abb8c167 could not be found. [ 700.575931] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 700.576083] env[61867]: INFO nova.compute.manager [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Took 0.07 seconds to destroy the instance on the hypervisor. [ 700.576319] env[61867]: DEBUG oslo.service.loopingcall [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.578496] env[61867]: DEBUG nova.compute.manager [-] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 700.578598] env[61867]: DEBUG nova.network.neutron [-] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 700.604760] env[61867]: DEBUG nova.network.neutron [-] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.615731] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.707838] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.764511] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0fb2d7-32d9-4404-a2cc-33030166aca0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.772375] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be96485-7772-409f-a010-ca5378276d2d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.802476] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3de28ad-e75a-4350-a8a0-e40507af1fd2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.809486] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c2c8473-6e79-4367-8512-93cac0701766 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.823224] env[61867]: DEBUG nova.compute.provider_tree [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.014773] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Releasing lock "refresh_cache-5c250675-3708-4c25-b370-6707c457357a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.015068] env[61867]: DEBUG nova.compute.manager [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 701.015328] env[61867]: DEBUG nova.compute.manager [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 701.015447] env[61867]: DEBUG nova.network.neutron [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 701.032271] env[61867]: DEBUG nova.network.neutron [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.107930] env[61867]: DEBUG nova.network.neutron [-] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.210308] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Releasing lock "refresh_cache-3f1696b0-4de4-4128-bc83-e539e48dc8e3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.210705] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 701.210912] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 701.211233] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ed6258a6-d812-4298-b4ac-23ed14a6f1a7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.221341] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e56199de-a8e9-4671-ad16-16a94cc1446e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.244869] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3f1696b0-4de4-4128-bc83-e539e48dc8e3 could not be found. [ 701.245146] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 701.245360] env[61867]: INFO nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Took 0.03 seconds to destroy the instance on the hypervisor. [ 701.245632] env[61867]: DEBUG oslo.service.loopingcall [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 701.246226] env[61867]: DEBUG nova.compute.manager [-] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 701.246343] env[61867]: DEBUG nova.network.neutron [-] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 701.262693] env[61867]: DEBUG nova.network.neutron [-] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.268269] env[61867]: DEBUG nova.compute.manager [req-b838fc6f-5d1e-47aa-a14c-12eed247f7bc req-448f9a5e-ee4c-43f6-a407-e87d6a628247 service nova] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Received event network-changed-0ccf49d0-6d5d-491b-a98a-d0afda9b598c {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 701.268490] env[61867]: DEBUG nova.compute.manager [req-b838fc6f-5d1e-47aa-a14c-12eed247f7bc req-448f9a5e-ee4c-43f6-a407-e87d6a628247 service nova] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Refreshing instance network info cache due to event network-changed-0ccf49d0-6d5d-491b-a98a-d0afda9b598c. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 701.268650] env[61867]: DEBUG oslo_concurrency.lockutils [req-b838fc6f-5d1e-47aa-a14c-12eed247f7bc req-448f9a5e-ee4c-43f6-a407-e87d6a628247 service nova] Acquiring lock "refresh_cache-3f1696b0-4de4-4128-bc83-e539e48dc8e3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.268878] env[61867]: DEBUG oslo_concurrency.lockutils [req-b838fc6f-5d1e-47aa-a14c-12eed247f7bc req-448f9a5e-ee4c-43f6-a407-e87d6a628247 service nova] Acquired lock "refresh_cache-3f1696b0-4de4-4128-bc83-e539e48dc8e3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.269321] env[61867]: DEBUG nova.network.neutron [req-b838fc6f-5d1e-47aa-a14c-12eed247f7bc req-448f9a5e-ee4c-43f6-a407-e87d6a628247 service nova] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Refreshing network info cache for port 0ccf49d0-6d5d-491b-a98a-d0afda9b598c {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 701.325771] env[61867]: DEBUG nova.scheduler.client.report [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.535815] env[61867]: DEBUG nova.network.neutron [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.609776] env[61867]: INFO nova.compute.manager [-] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Took 1.03 seconds to deallocate network for instance. [ 701.766820] env[61867]: DEBUG nova.network.neutron [-] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.787107] env[61867]: DEBUG nova.network.neutron [req-b838fc6f-5d1e-47aa-a14c-12eed247f7bc req-448f9a5e-ee4c-43f6-a407-e87d6a628247 service nova] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.830780] env[61867]: DEBUG oslo_concurrency.lockutils [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.921s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.831550] env[61867]: ERROR nova.compute.manager [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port feaf8561-0b7e-4022-ba56-899778e4aaf7, please check neutron logs for more information. [ 701.831550] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Traceback (most recent call last): [ 701.831550] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.831550] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] self.driver.spawn(context, instance, image_meta, [ 701.831550] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 701.831550] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.831550] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.831550] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] vm_ref = self.build_virtual_machine(instance, [ 701.831550] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.831550] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.831550] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.831882] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] for vif in network_info: [ 701.831882] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.831882] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] return self._sync_wrapper(fn, *args, **kwargs) [ 701.831882] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.831882] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] self.wait() [ 701.831882] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.831882] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] self[:] = self._gt.wait() [ 701.831882] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.831882] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] return self._exit_event.wait() [ 701.831882] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 701.831882] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] result = hub.switch() [ 701.831882] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 701.831882] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] return self.greenlet.switch() [ 701.832395] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.832395] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] result = function(*args, **kwargs) [ 701.832395] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.832395] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] return func(*args, **kwargs) [ 701.832395] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.832395] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] raise e [ 701.832395] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.832395] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] nwinfo = self.network_api.allocate_for_instance( [ 701.832395] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.832395] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] created_port_ids = self._update_ports_for_instance( [ 701.832395] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.832395] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] with excutils.save_and_reraise_exception(): [ 701.832395] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.832782] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] self.force_reraise() [ 701.832782] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.832782] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] raise self.value [ 701.832782] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.832782] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] updated_port = self._update_port( [ 701.832782] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.832782] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] _ensure_no_port_binding_failure(port) [ 701.832782] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.832782] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] raise exception.PortBindingFailed(port_id=port['id']) [ 701.832782] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] nova.exception.PortBindingFailed: Binding failed for port feaf8561-0b7e-4022-ba56-899778e4aaf7, please check neutron logs for more information. [ 701.832782] env[61867]: ERROR nova.compute.manager [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] [ 701.833134] env[61867]: DEBUG nova.compute.utils [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Binding failed for port feaf8561-0b7e-4022-ba56-899778e4aaf7, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 701.833484] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.762s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.834983] env[61867]: INFO nova.compute.claims [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.838686] env[61867]: DEBUG nova.compute.manager [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Build of instance 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d was re-scheduled: Binding failed for port feaf8561-0b7e-4022-ba56-899778e4aaf7, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 701.839229] env[61867]: DEBUG nova.compute.manager [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 701.839472] env[61867]: DEBUG oslo_concurrency.lockutils [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "refresh_cache-4d9c1c85-9d49-4c7c-bcac-f22f5842db0d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.839707] env[61867]: DEBUG oslo_concurrency.lockutils [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquired lock "refresh_cache-4d9c1c85-9d49-4c7c-bcac-f22f5842db0d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.840014] env[61867]: DEBUG nova.network.neutron [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 701.854794] env[61867]: DEBUG nova.network.neutron [req-b838fc6f-5d1e-47aa-a14c-12eed247f7bc req-448f9a5e-ee4c-43f6-a407-e87d6a628247 service nova] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.038543] env[61867]: INFO nova.compute.manager [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 5c250675-3708-4c25-b370-6707c457357a] Took 1.02 seconds to deallocate network for instance. [ 702.163382] env[61867]: INFO nova.compute.manager [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Took 0.55 seconds to detach 1 volumes for instance. [ 702.165680] env[61867]: DEBUG nova.compute.claims [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 702.165859] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.271027] env[61867]: INFO nova.compute.manager [-] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Took 1.02 seconds to deallocate network for instance. [ 702.274155] env[61867]: DEBUG nova.compute.claims [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 702.274155] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.356882] env[61867]: DEBUG nova.network.neutron [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.358767] env[61867]: DEBUG oslo_concurrency.lockutils [req-b838fc6f-5d1e-47aa-a14c-12eed247f7bc req-448f9a5e-ee4c-43f6-a407-e87d6a628247 service nova] Releasing lock "refresh_cache-3f1696b0-4de4-4128-bc83-e539e48dc8e3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.359015] env[61867]: DEBUG nova.compute.manager [req-b838fc6f-5d1e-47aa-a14c-12eed247f7bc req-448f9a5e-ee4c-43f6-a407-e87d6a628247 service nova] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Received event network-vif-deleted-0ccf49d0-6d5d-491b-a98a-d0afda9b598c {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 702.403127] env[61867]: DEBUG nova.network.neutron [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.905460] env[61867]: DEBUG oslo_concurrency.lockutils [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Releasing lock "refresh_cache-4d9c1c85-9d49-4c7c-bcac-f22f5842db0d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.905694] env[61867]: DEBUG nova.compute.manager [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 702.905861] env[61867]: DEBUG nova.compute.manager [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 702.906047] env[61867]: DEBUG nova.network.neutron [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 702.923696] env[61867]: DEBUG nova.network.neutron [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.069443] env[61867]: INFO nova.scheduler.client.report [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Deleted allocations for instance 5c250675-3708-4c25-b370-6707c457357a [ 703.155553] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94ddbc14-2a34-4537-aa53-e8ec4965874e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.162352] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b46d81f-f00a-4ca9-9467-06eb73a3cc0b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.191877] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9dbb917-0cd5-44a5-8913-81c334b31579 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.199136] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b6fdf31-d43a-4e77-9074-f9a3f348d1dc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.212091] env[61867]: DEBUG nova.compute.provider_tree [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.427046] env[61867]: DEBUG nova.network.neutron [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.581865] env[61867]: DEBUG oslo_concurrency.lockutils [None req-eba8f612-b75b-451d-81e0-30fc77935549 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "5c250675-3708-4c25-b370-6707c457357a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.780s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.717064] env[61867]: DEBUG nova.scheduler.client.report [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.929690] env[61867]: INFO nova.compute.manager [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d] Took 1.02 seconds to deallocate network for instance. [ 704.084554] env[61867]: DEBUG nova.compute.manager [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 704.219929] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.220486] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 704.223280] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.360s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.223280] env[61867]: DEBUG nova.objects.instance [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61867) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 704.605454] env[61867]: DEBUG oslo_concurrency.lockutils [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.727497] env[61867]: DEBUG nova.compute.utils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 704.728905] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 704.729095] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 704.779632] env[61867]: DEBUG nova.policy [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d7b6b0de3374b09afb54bbc5c7c2382', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a47120d2cb044db0bb45a9e801d1fe2d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 704.965380] env[61867]: INFO nova.scheduler.client.report [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Deleted allocations for instance 4d9c1c85-9d49-4c7c-bcac-f22f5842db0d [ 705.053040] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Successfully created port: fddc7c36-1b1e-4671-a7dd-fe475981f035 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 705.234775] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 705.239438] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bd4a5bd9-e2b2-4da9-9e6d-c4e6e0b9caa6 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.240574] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.288s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.243554] env[61867]: INFO nova.compute.claims [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 705.475193] env[61867]: DEBUG oslo_concurrency.lockutils [None req-273c25b8-4c0f-4ad9-9e3c-b17fd60aa78c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "4d9c1c85-9d49-4c7c-bcac-f22f5842db0d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.445s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.901204] env[61867]: DEBUG nova.compute.manager [req-df9ce383-7707-4651-ae28-20872d5c690a req-0ba402a3-fa28-46a9-b73f-ede9b8422916 service nova] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Received event network-changed-fddc7c36-1b1e-4671-a7dd-fe475981f035 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 705.901204] env[61867]: DEBUG nova.compute.manager [req-df9ce383-7707-4651-ae28-20872d5c690a req-0ba402a3-fa28-46a9-b73f-ede9b8422916 service nova] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Refreshing instance network info cache due to event network-changed-fddc7c36-1b1e-4671-a7dd-fe475981f035. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 705.901204] env[61867]: DEBUG oslo_concurrency.lockutils [req-df9ce383-7707-4651-ae28-20872d5c690a req-0ba402a3-fa28-46a9-b73f-ede9b8422916 service nova] Acquiring lock "refresh_cache-2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.901204] env[61867]: DEBUG oslo_concurrency.lockutils [req-df9ce383-7707-4651-ae28-20872d5c690a req-0ba402a3-fa28-46a9-b73f-ede9b8422916 service nova] Acquired lock "refresh_cache-2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.901577] env[61867]: DEBUG nova.network.neutron [req-df9ce383-7707-4651-ae28-20872d5c690a req-0ba402a3-fa28-46a9-b73f-ede9b8422916 service nova] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Refreshing network info cache for port fddc7c36-1b1e-4671-a7dd-fe475981f035 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 705.936611] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "1df8427c-e75d-4b60-a92a-b5ba76b67081" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.936611] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "1df8427c-e75d-4b60-a92a-b5ba76b67081" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.978080] env[61867]: DEBUG nova.compute.manager [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 706.068555] env[61867]: ERROR nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fddc7c36-1b1e-4671-a7dd-fe475981f035, please check neutron logs for more information. [ 706.068555] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 706.068555] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.068555] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 706.068555] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.068555] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 706.068555] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.068555] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 706.068555] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.068555] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 706.068555] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.068555] env[61867]: ERROR nova.compute.manager raise self.value [ 706.068555] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.068555] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 706.068555] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.068555] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 706.069085] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.069085] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 706.069085] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fddc7c36-1b1e-4671-a7dd-fe475981f035, please check neutron logs for more information. [ 706.069085] env[61867]: ERROR nova.compute.manager [ 706.069085] env[61867]: Traceback (most recent call last): [ 706.069085] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 706.069085] env[61867]: listener.cb(fileno) [ 706.069085] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.069085] env[61867]: result = function(*args, **kwargs) [ 706.069085] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.069085] env[61867]: return func(*args, **kwargs) [ 706.069085] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.069085] env[61867]: raise e [ 706.069085] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.069085] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 706.069085] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.069085] env[61867]: created_port_ids = self._update_ports_for_instance( [ 706.069085] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.069085] env[61867]: with excutils.save_and_reraise_exception(): [ 706.069085] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.069085] env[61867]: self.force_reraise() [ 706.069085] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.069085] env[61867]: raise self.value [ 706.069085] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.069085] env[61867]: updated_port = self._update_port( [ 706.069085] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.069085] env[61867]: _ensure_no_port_binding_failure(port) [ 706.069085] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.069085] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 706.070026] env[61867]: nova.exception.PortBindingFailed: Binding failed for port fddc7c36-1b1e-4671-a7dd-fe475981f035, please check neutron logs for more information. [ 706.070026] env[61867]: Removing descriptor: 21 [ 706.246930] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 706.294089] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 706.294335] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 706.294490] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 706.294664] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 706.294807] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 706.294972] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 706.295168] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 706.295318] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 706.295478] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 706.295646] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 706.295822] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 706.297357] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ed806a-4db5-4cb9-8dd4-a19c5a0c2e0a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.307588] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de87cc76-245e-4f6d-88c1-fc24e307a330 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.321509] env[61867]: ERROR nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fddc7c36-1b1e-4671-a7dd-fe475981f035, please check neutron logs for more information. [ 706.321509] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Traceback (most recent call last): [ 706.321509] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 706.321509] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] yield resources [ 706.321509] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 706.321509] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] self.driver.spawn(context, instance, image_meta, [ 706.321509] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 706.321509] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.321509] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.321509] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] vm_ref = self.build_virtual_machine(instance, [ 706.321509] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.321965] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.321965] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.321965] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] for vif in network_info: [ 706.321965] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 706.321965] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] return self._sync_wrapper(fn, *args, **kwargs) [ 706.321965] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 706.321965] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] self.wait() [ 706.321965] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 706.321965] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] self[:] = self._gt.wait() [ 706.321965] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.321965] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] return self._exit_event.wait() [ 706.321965] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 706.321965] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] current.throw(*self._exc) [ 706.322418] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.322418] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] result = function(*args, **kwargs) [ 706.322418] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.322418] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] return func(*args, **kwargs) [ 706.322418] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.322418] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] raise e [ 706.322418] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.322418] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] nwinfo = self.network_api.allocate_for_instance( [ 706.322418] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.322418] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] created_port_ids = self._update_ports_for_instance( [ 706.322418] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.322418] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] with excutils.save_and_reraise_exception(): [ 706.322418] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.322844] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] self.force_reraise() [ 706.322844] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.322844] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] raise self.value [ 706.322844] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.322844] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] updated_port = self._update_port( [ 706.322844] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.322844] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] _ensure_no_port_binding_failure(port) [ 706.322844] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.322844] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] raise exception.PortBindingFailed(port_id=port['id']) [ 706.322844] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] nova.exception.PortBindingFailed: Binding failed for port fddc7c36-1b1e-4671-a7dd-fe475981f035, please check neutron logs for more information. [ 706.322844] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] [ 706.322844] env[61867]: INFO nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Terminating instance [ 706.325866] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquiring lock "refresh_cache-2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.424573] env[61867]: DEBUG nova.network.neutron [req-df9ce383-7707-4651-ae28-20872d5c690a req-0ba402a3-fa28-46a9-b73f-ede9b8422916 service nova] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.486714] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "81e70c36-04ea-450c-9383-53ef069d1c46" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.487042] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "81e70c36-04ea-450c-9383-53ef069d1c46" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.500365] env[61867]: DEBUG oslo_concurrency.lockutils [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.523946] env[61867]: DEBUG nova.network.neutron [req-df9ce383-7707-4651-ae28-20872d5c690a req-0ba402a3-fa28-46a9-b73f-ede9b8422916 service nova] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.635947] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a97ef36-4070-45c1-9dd2-e06b636c4d42 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.643850] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e8a9df3-4f69-4beb-a5ce-fbabc3483d97 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.675056] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b472f4f-7de1-4705-92a3-0596d68619cc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.682018] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-870fcb97-166b-4586-abb5-33373683dbae {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.694749] env[61867]: DEBUG nova.compute.provider_tree [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.027598] env[61867]: DEBUG oslo_concurrency.lockutils [req-df9ce383-7707-4651-ae28-20872d5c690a req-0ba402a3-fa28-46a9-b73f-ede9b8422916 service nova] Releasing lock "refresh_cache-2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.027598] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquired lock "refresh_cache-2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.027598] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.197702] env[61867]: DEBUG nova.scheduler.client.report [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.546043] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.619992] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.702426] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.702959] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 707.705437] env[61867]: DEBUG oslo_concurrency.lockutils [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.060s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.922816] env[61867]: DEBUG nova.compute.manager [req-8e78096d-f308-4d13-a169-c87e5d8086e5 req-d777a99c-efb8-47be-a86c-19fd7f5e62fc service nova] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Received event network-vif-deleted-fddc7c36-1b1e-4671-a7dd-fe475981f035 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 708.122417] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Releasing lock "refresh_cache-2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.122841] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 708.123048] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 708.123351] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3d89c61b-22ff-426b-ae6a-c1bae2eaa76f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.132738] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-117d4c40-3fab-48ca-8966-d53561913140 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.154171] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de could not be found. [ 708.154395] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 708.154579] env[61867]: INFO nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Took 0.03 seconds to destroy the instance on the hypervisor. [ 708.154813] env[61867]: DEBUG oslo.service.loopingcall [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.155027] env[61867]: DEBUG nova.compute.manager [-] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.155128] env[61867]: DEBUG nova.network.neutron [-] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 708.169677] env[61867]: DEBUG nova.network.neutron [-] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.209893] env[61867]: DEBUG nova.compute.utils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 708.211201] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 708.211648] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 708.255408] env[61867]: DEBUG nova.policy [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d7b6b0de3374b09afb54bbc5c7c2382', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a47120d2cb044db0bb45a9e801d1fe2d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 708.513534] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Successfully created port: b13b1e49-c2b5-4a52-adec-93dda37f48af {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 708.539529] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6d932c-d339-4e0b-b335-c4e04055cdb5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.549182] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361a0f21-175e-4232-861e-7fb1a4562594 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.579336] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d6d0cf-a902-4295-aa53-26fbeb13adc2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.586731] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46fdea8d-b211-4cd2-93bf-2d4f6fd08d23 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.601320] env[61867]: DEBUG nova.compute.provider_tree [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.673937] env[61867]: DEBUG nova.network.neutron [-] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.716628] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 709.107586] env[61867]: DEBUG nova.scheduler.client.report [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 709.176636] env[61867]: INFO nova.compute.manager [-] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Took 1.02 seconds to deallocate network for instance. [ 709.179012] env[61867]: DEBUG nova.compute.claims [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 709.179211] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.296748] env[61867]: ERROR nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b13b1e49-c2b5-4a52-adec-93dda37f48af, please check neutron logs for more information. [ 709.296748] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 709.296748] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.296748] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 709.296748] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.296748] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 709.296748] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.296748] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 709.296748] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.296748] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 709.296748] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.296748] env[61867]: ERROR nova.compute.manager raise self.value [ 709.296748] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.296748] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 709.296748] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.296748] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 709.297681] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.297681] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 709.297681] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b13b1e49-c2b5-4a52-adec-93dda37f48af, please check neutron logs for more information. [ 709.297681] env[61867]: ERROR nova.compute.manager [ 709.297681] env[61867]: Traceback (most recent call last): [ 709.297681] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 709.297681] env[61867]: listener.cb(fileno) [ 709.297681] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.297681] env[61867]: result = function(*args, **kwargs) [ 709.297681] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 709.297681] env[61867]: return func(*args, **kwargs) [ 709.297681] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 709.297681] env[61867]: raise e [ 709.297681] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.297681] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 709.297681] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.297681] env[61867]: created_port_ids = self._update_ports_for_instance( [ 709.297681] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.297681] env[61867]: with excutils.save_and_reraise_exception(): [ 709.297681] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.297681] env[61867]: self.force_reraise() [ 709.297681] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.297681] env[61867]: raise self.value [ 709.297681] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.297681] env[61867]: updated_port = self._update_port( [ 709.297681] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.297681] env[61867]: _ensure_no_port_binding_failure(port) [ 709.297681] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.297681] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 709.298482] env[61867]: nova.exception.PortBindingFailed: Binding failed for port b13b1e49-c2b5-4a52-adec-93dda37f48af, please check neutron logs for more information. [ 709.298482] env[61867]: Removing descriptor: 21 [ 709.614726] env[61867]: DEBUG oslo_concurrency.lockutils [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.909s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.615385] env[61867]: ERROR nova.compute.manager [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 78d19f74-61d7-47cc-bafd-33d2683cb241, please check neutron logs for more information. [ 709.615385] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Traceback (most recent call last): [ 709.615385] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 709.615385] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] self.driver.spawn(context, instance, image_meta, [ 709.615385] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 709.615385] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] self._vmops.spawn(context, instance, image_meta, injected_files, [ 709.615385] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 709.615385] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] vm_ref = self.build_virtual_machine(instance, [ 709.615385] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 709.615385] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] vif_infos = vmwarevif.get_vif_info(self._session, [ 709.615385] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 709.615715] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] for vif in network_info: [ 709.615715] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 709.615715] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] return self._sync_wrapper(fn, *args, **kwargs) [ 709.615715] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 709.615715] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] self.wait() [ 709.615715] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 709.615715] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] self[:] = self._gt.wait() [ 709.615715] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 709.615715] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] return self._exit_event.wait() [ 709.615715] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 709.615715] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] current.throw(*self._exc) [ 709.615715] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.615715] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] result = function(*args, **kwargs) [ 709.616087] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 709.616087] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] return func(*args, **kwargs) [ 709.616087] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 709.616087] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] raise e [ 709.616087] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.616087] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] nwinfo = self.network_api.allocate_for_instance( [ 709.616087] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.616087] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] created_port_ids = self._update_ports_for_instance( [ 709.616087] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.616087] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] with excutils.save_and_reraise_exception(): [ 709.616087] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.616087] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] self.force_reraise() [ 709.616087] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.616491] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] raise self.value [ 709.616491] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.616491] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] updated_port = self._update_port( [ 709.616491] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.616491] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] _ensure_no_port_binding_failure(port) [ 709.616491] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.616491] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] raise exception.PortBindingFailed(port_id=port['id']) [ 709.616491] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] nova.exception.PortBindingFailed: Binding failed for port 78d19f74-61d7-47cc-bafd-33d2683cb241, please check neutron logs for more information. [ 709.616491] env[61867]: ERROR nova.compute.manager [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] [ 709.616491] env[61867]: DEBUG nova.compute.utils [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Binding failed for port 78d19f74-61d7-47cc-bafd-33d2683cb241, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 709.617392] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.159s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.618835] env[61867]: INFO nova.compute.claims [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 709.621685] env[61867]: DEBUG nova.compute.manager [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Build of instance 59d2ee3a-756f-4dda-a70a-6d43aaf32049 was re-scheduled: Binding failed for port 78d19f74-61d7-47cc-bafd-33d2683cb241, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 709.622022] env[61867]: DEBUG nova.compute.manager [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 709.622258] env[61867]: DEBUG oslo_concurrency.lockutils [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Acquiring lock "refresh_cache-59d2ee3a-756f-4dda-a70a-6d43aaf32049" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.622406] env[61867]: DEBUG oslo_concurrency.lockutils [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Acquired lock "refresh_cache-59d2ee3a-756f-4dda-a70a-6d43aaf32049" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.622594] env[61867]: DEBUG nova.network.neutron [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 709.727038] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 709.752341] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 709.752583] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 709.752736] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 709.752926] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 709.753098] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 709.753245] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 709.753445] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 709.753598] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 709.753756] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 709.753910] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 709.754100] env[61867]: DEBUG nova.virt.hardware [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 709.754933] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b85a5707-78fc-4537-85c8-c50cea0d1cd4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.762705] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ba0779-fa58-4b92-a64e-1c905529ba02 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.776094] env[61867]: ERROR nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b13b1e49-c2b5-4a52-adec-93dda37f48af, please check neutron logs for more information. [ 709.776094] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Traceback (most recent call last): [ 709.776094] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 709.776094] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] yield resources [ 709.776094] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 709.776094] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] self.driver.spawn(context, instance, image_meta, [ 709.776094] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 709.776094] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 709.776094] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 709.776094] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] vm_ref = self.build_virtual_machine(instance, [ 709.776094] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 709.776529] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] vif_infos = vmwarevif.get_vif_info(self._session, [ 709.776529] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 709.776529] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] for vif in network_info: [ 709.776529] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 709.776529] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] return self._sync_wrapper(fn, *args, **kwargs) [ 709.776529] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 709.776529] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] self.wait() [ 709.776529] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 709.776529] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] self[:] = self._gt.wait() [ 709.776529] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 709.776529] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] return self._exit_event.wait() [ 709.776529] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 709.776529] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] current.throw(*self._exc) [ 709.777167] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.777167] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] result = function(*args, **kwargs) [ 709.777167] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 709.777167] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] return func(*args, **kwargs) [ 709.777167] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 709.777167] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] raise e [ 709.777167] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.777167] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] nwinfo = self.network_api.allocate_for_instance( [ 709.777167] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.777167] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] created_port_ids = self._update_ports_for_instance( [ 709.777167] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.777167] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] with excutils.save_and_reraise_exception(): [ 709.777167] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.777610] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] self.force_reraise() [ 709.777610] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.777610] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] raise self.value [ 709.777610] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.777610] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] updated_port = self._update_port( [ 709.777610] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.777610] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] _ensure_no_port_binding_failure(port) [ 709.777610] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.777610] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] raise exception.PortBindingFailed(port_id=port['id']) [ 709.777610] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] nova.exception.PortBindingFailed: Binding failed for port b13b1e49-c2b5-4a52-adec-93dda37f48af, please check neutron logs for more information. [ 709.777610] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] [ 709.777610] env[61867]: INFO nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Terminating instance [ 709.778915] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquiring lock "refresh_cache-add3bbbb-fd30-49ad-b95e-601af5d790e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.779854] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquired lock "refresh_cache-add3bbbb-fd30-49ad-b95e-601af5d790e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.779854] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 709.948156] env[61867]: DEBUG nova.compute.manager [req-1f20cb9f-5f4c-4ace-a440-6edd90b83a51 req-c7aa7e19-71f5-4c6f-8d7b-514609678d1e service nova] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Received event network-changed-b13b1e49-c2b5-4a52-adec-93dda37f48af {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 709.948361] env[61867]: DEBUG nova.compute.manager [req-1f20cb9f-5f4c-4ace-a440-6edd90b83a51 req-c7aa7e19-71f5-4c6f-8d7b-514609678d1e service nova] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Refreshing instance network info cache due to event network-changed-b13b1e49-c2b5-4a52-adec-93dda37f48af. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 709.948545] env[61867]: DEBUG oslo_concurrency.lockutils [req-1f20cb9f-5f4c-4ace-a440-6edd90b83a51 req-c7aa7e19-71f5-4c6f-8d7b-514609678d1e service nova] Acquiring lock "refresh_cache-add3bbbb-fd30-49ad-b95e-601af5d790e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.143181] env[61867]: DEBUG nova.network.neutron [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.224377] env[61867]: DEBUG nova.network.neutron [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.299108] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.381435] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.726692] env[61867]: DEBUG oslo_concurrency.lockutils [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Releasing lock "refresh_cache-59d2ee3a-756f-4dda-a70a-6d43aaf32049" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.727000] env[61867]: DEBUG nova.compute.manager [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 710.727184] env[61867]: DEBUG nova.compute.manager [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 710.727351] env[61867]: DEBUG nova.network.neutron [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 710.741807] env[61867]: DEBUG nova.network.neutron [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.815350] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 710.815719] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 710.887260] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Releasing lock "refresh_cache-add3bbbb-fd30-49ad-b95e-601af5d790e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.887260] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 710.887260] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 710.887260] env[61867]: DEBUG oslo_concurrency.lockutils [req-1f20cb9f-5f4c-4ace-a440-6edd90b83a51 req-c7aa7e19-71f5-4c6f-8d7b-514609678d1e service nova] Acquired lock "refresh_cache-add3bbbb-fd30-49ad-b95e-601af5d790e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.887260] env[61867]: DEBUG nova.network.neutron [req-1f20cb9f-5f4c-4ace-a440-6edd90b83a51 req-c7aa7e19-71f5-4c6f-8d7b-514609678d1e service nova] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Refreshing network info cache for port b13b1e49-c2b5-4a52-adec-93dda37f48af {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 710.887514] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ada88b4b-1df0-4f3d-a927-29f1af07a5e6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.894944] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aad95e4-1f7a-49b7-8153-a877ec477113 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.918303] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance add3bbbb-fd30-49ad-b95e-601af5d790e4 could not be found. [ 710.918556] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 710.918737] env[61867]: INFO nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 710.919012] env[61867]: DEBUG oslo.service.loopingcall [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 710.921554] env[61867]: DEBUG nova.compute.manager [-] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 710.921652] env[61867]: DEBUG nova.network.neutron [-] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 710.937409] env[61867]: DEBUG nova.network.neutron [-] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.944949] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-697ffc51-49f8-4abf-ab55-4ddeea298d40 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.954633] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e25c5d8-8486-40d0-9997-22fd2c5dbaa1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.983647] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa4b914-43f5-4fb1-81e4-d0ed7dd5cc5e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.990731] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76975ab1-a27b-4994-b288-1d0862220a48 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.003569] env[61867]: DEBUG nova.compute.provider_tree [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.244449] env[61867]: DEBUG nova.network.neutron [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.322922] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 711.323040] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Starting heal instance info cache {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 711.323131] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Rebuilding the list of instances to heal {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 711.403133] env[61867]: DEBUG nova.network.neutron [req-1f20cb9f-5f4c-4ace-a440-6edd90b83a51 req-c7aa7e19-71f5-4c6f-8d7b-514609678d1e service nova] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.440170] env[61867]: DEBUG nova.network.neutron [-] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.472985] env[61867]: DEBUG nova.network.neutron [req-1f20cb9f-5f4c-4ace-a440-6edd90b83a51 req-c7aa7e19-71f5-4c6f-8d7b-514609678d1e service nova] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.506821] env[61867]: DEBUG nova.scheduler.client.report [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 711.750429] env[61867]: INFO nova.compute.manager [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] [instance: 59d2ee3a-756f-4dda-a70a-6d43aaf32049] Took 1.02 seconds to deallocate network for instance. [ 711.828827] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 774e7c38-426a-441d-8252-3daa424be558] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 711.829028] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 711.829167] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 711.829325] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 711.829471] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 711.829598] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 711.857496] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "refresh_cache-37ed9e18-8dba-459d-bc67-e3ce3f9cdaea" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.857641] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquired lock "refresh_cache-37ed9e18-8dba-459d-bc67-e3ce3f9cdaea" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.857784] env[61867]: DEBUG nova.network.neutron [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Forcefully refreshing network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 711.857936] env[61867]: DEBUG nova.objects.instance [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lazy-loading 'info_cache' on Instance uuid 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 711.942384] env[61867]: INFO nova.compute.manager [-] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Took 1.02 seconds to deallocate network for instance. [ 711.944800] env[61867]: DEBUG nova.compute.claims [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 711.945012] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.975725] env[61867]: DEBUG oslo_concurrency.lockutils [req-1f20cb9f-5f4c-4ace-a440-6edd90b83a51 req-c7aa7e19-71f5-4c6f-8d7b-514609678d1e service nova] Releasing lock "refresh_cache-add3bbbb-fd30-49ad-b95e-601af5d790e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.976093] env[61867]: DEBUG nova.compute.manager [req-1f20cb9f-5f4c-4ace-a440-6edd90b83a51 req-c7aa7e19-71f5-4c6f-8d7b-514609678d1e service nova] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Received event network-vif-deleted-b13b1e49-c2b5-4a52-adec-93dda37f48af {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 712.010920] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.011440] env[61867]: DEBUG nova.compute.manager [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 712.013877] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.468s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.014112] env[61867]: DEBUG nova.objects.instance [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Lazy-loading 'resources' on Instance uuid 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 712.520105] env[61867]: DEBUG nova.compute.utils [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 712.520886] env[61867]: DEBUG nova.compute.manager [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 712.521068] env[61867]: DEBUG nova.network.neutron [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 712.568171] env[61867]: DEBUG nova.policy [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ea5fdaf720f4d0e91be136e70cad82c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '54d2684558bc48f0a3cd78ab3ef667dc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 712.776060] env[61867]: INFO nova.scheduler.client.report [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Deleted allocations for instance 59d2ee3a-756f-4dda-a70a-6d43aaf32049 [ 712.800655] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45ba72e8-053e-4a28-b4c9-22c5d5c5112f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.809015] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f4fb07-9fff-41cd-89ac-0dd8ed73a1f9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.842241] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-327dda29-189b-4d03-9c20-42c245140c6e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.848700] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92a0df8-42c2-40da-9269-3fab46317fa7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.865340] env[61867]: DEBUG nova.compute.provider_tree [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.885436] env[61867]: DEBUG nova.network.neutron [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.911216] env[61867]: DEBUG nova.network.neutron [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Successfully created port: e2d64557-ec03-4212-bfcc-6e3d8cc9f0f0 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 713.023724] env[61867]: DEBUG nova.compute.manager [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 713.285500] env[61867]: DEBUG oslo_concurrency.lockutils [None req-acc73891-2aa6-4230-9510-1a05586c274e tempest-ServerAddressesTestJSON-470789647 tempest-ServerAddressesTestJSON-470789647-project-member] Lock "59d2ee3a-756f-4dda-a70a-6d43aaf32049" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.839s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.369413] env[61867]: DEBUG nova.scheduler.client.report [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 713.475115] env[61867]: DEBUG nova.network.neutron [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.638615] env[61867]: DEBUG nova.compute.manager [req-c774b874-4e92-4d07-a56d-94c18dd78afe req-d0dd7296-246d-4e03-a112-4d5022617e1d service nova] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Received event network-changed-e2d64557-ec03-4212-bfcc-6e3d8cc9f0f0 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 713.638708] env[61867]: DEBUG nova.compute.manager [req-c774b874-4e92-4d07-a56d-94c18dd78afe req-d0dd7296-246d-4e03-a112-4d5022617e1d service nova] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Refreshing instance network info cache due to event network-changed-e2d64557-ec03-4212-bfcc-6e3d8cc9f0f0. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 713.638870] env[61867]: DEBUG oslo_concurrency.lockutils [req-c774b874-4e92-4d07-a56d-94c18dd78afe req-d0dd7296-246d-4e03-a112-4d5022617e1d service nova] Acquiring lock "refresh_cache-5dab4fcf-1a5e-493a-842b-b652e99d67ba" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.639100] env[61867]: DEBUG oslo_concurrency.lockutils [req-c774b874-4e92-4d07-a56d-94c18dd78afe req-d0dd7296-246d-4e03-a112-4d5022617e1d service nova] Acquired lock "refresh_cache-5dab4fcf-1a5e-493a-842b-b652e99d67ba" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.639264] env[61867]: DEBUG nova.network.neutron [req-c774b874-4e92-4d07-a56d-94c18dd78afe req-d0dd7296-246d-4e03-a112-4d5022617e1d service nova] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Refreshing network info cache for port e2d64557-ec03-4212-bfcc-6e3d8cc9f0f0 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 713.781191] env[61867]: ERROR nova.compute.manager [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e2d64557-ec03-4212-bfcc-6e3d8cc9f0f0, please check neutron logs for more information. [ 713.781191] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 713.781191] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.781191] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 713.781191] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.781191] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 713.781191] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.781191] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 713.781191] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.781191] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 713.781191] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.781191] env[61867]: ERROR nova.compute.manager raise self.value [ 713.781191] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.781191] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 713.781191] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.781191] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 713.781898] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.781898] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 713.781898] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e2d64557-ec03-4212-bfcc-6e3d8cc9f0f0, please check neutron logs for more information. [ 713.781898] env[61867]: ERROR nova.compute.manager [ 713.781898] env[61867]: Traceback (most recent call last): [ 713.781898] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 713.781898] env[61867]: listener.cb(fileno) [ 713.781898] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.781898] env[61867]: result = function(*args, **kwargs) [ 713.781898] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 713.781898] env[61867]: return func(*args, **kwargs) [ 713.781898] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 713.781898] env[61867]: raise e [ 713.781898] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.781898] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 713.781898] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.781898] env[61867]: created_port_ids = self._update_ports_for_instance( [ 713.781898] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.781898] env[61867]: with excutils.save_and_reraise_exception(): [ 713.781898] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.781898] env[61867]: self.force_reraise() [ 713.781898] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.781898] env[61867]: raise self.value [ 713.781898] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.781898] env[61867]: updated_port = self._update_port( [ 713.781898] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.781898] env[61867]: _ensure_no_port_binding_failure(port) [ 713.781898] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.781898] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 713.782845] env[61867]: nova.exception.PortBindingFailed: Binding failed for port e2d64557-ec03-4212-bfcc-6e3d8cc9f0f0, please check neutron logs for more information. [ 713.782845] env[61867]: Removing descriptor: 21 [ 713.787833] env[61867]: DEBUG nova.compute.manager [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 713.873943] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.860s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.876388] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.461s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.898921] env[61867]: INFO nova.scheduler.client.report [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Deleted allocations for instance 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea [ 713.976124] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Releasing lock "refresh_cache-37ed9e18-8dba-459d-bc67-e3ce3f9cdaea" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.976345] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Updated the network info_cache for instance {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 713.976537] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 713.976766] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 713.976826] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 713.976974] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 713.977140] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 713.977302] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 713.977426] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61867) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 713.977567] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 714.032831] env[61867]: DEBUG nova.compute.manager [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 714.062269] env[61867]: DEBUG nova.virt.hardware [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 714.062512] env[61867]: DEBUG nova.virt.hardware [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 714.062666] env[61867]: DEBUG nova.virt.hardware [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 714.062843] env[61867]: DEBUG nova.virt.hardware [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 714.062989] env[61867]: DEBUG nova.virt.hardware [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 714.063150] env[61867]: DEBUG nova.virt.hardware [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 714.063357] env[61867]: DEBUG nova.virt.hardware [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 714.063515] env[61867]: DEBUG nova.virt.hardware [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 714.063992] env[61867]: DEBUG nova.virt.hardware [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 714.063992] env[61867]: DEBUG nova.virt.hardware [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 714.063992] env[61867]: DEBUG nova.virt.hardware [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 714.064838] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c3ef37-a482-478e-89f5-6ffba91b492c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.076240] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf9e5091-890a-48b7-9ec1-c6e34abd4062 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.090335] env[61867]: ERROR nova.compute.manager [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e2d64557-ec03-4212-bfcc-6e3d8cc9f0f0, please check neutron logs for more information. [ 714.090335] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Traceback (most recent call last): [ 714.090335] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 714.090335] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] yield resources [ 714.090335] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 714.090335] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] self.driver.spawn(context, instance, image_meta, [ 714.090335] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 714.090335] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.090335] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.090335] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] vm_ref = self.build_virtual_machine(instance, [ 714.090335] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.090701] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.090701] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.090701] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] for vif in network_info: [ 714.090701] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 714.090701] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] return self._sync_wrapper(fn, *args, **kwargs) [ 714.090701] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 714.090701] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] self.wait() [ 714.090701] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 714.090701] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] self[:] = self._gt.wait() [ 714.090701] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.090701] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] return self._exit_event.wait() [ 714.090701] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 714.090701] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] current.throw(*self._exc) [ 714.091069] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.091069] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] result = function(*args, **kwargs) [ 714.091069] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 714.091069] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] return func(*args, **kwargs) [ 714.091069] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.091069] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] raise e [ 714.091069] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.091069] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] nwinfo = self.network_api.allocate_for_instance( [ 714.091069] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 714.091069] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] created_port_ids = self._update_ports_for_instance( [ 714.091069] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 714.091069] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] with excutils.save_and_reraise_exception(): [ 714.091069] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.091478] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] self.force_reraise() [ 714.091478] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.091478] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] raise self.value [ 714.091478] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 714.091478] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] updated_port = self._update_port( [ 714.091478] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.091478] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] _ensure_no_port_binding_failure(port) [ 714.091478] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.091478] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] raise exception.PortBindingFailed(port_id=port['id']) [ 714.091478] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] nova.exception.PortBindingFailed: Binding failed for port e2d64557-ec03-4212-bfcc-6e3d8cc9f0f0, please check neutron logs for more information. [ 714.091478] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] [ 714.091478] env[61867]: INFO nova.compute.manager [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Terminating instance [ 714.092889] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Acquiring lock "refresh_cache-5dab4fcf-1a5e-493a-842b-b652e99d67ba" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.160260] env[61867]: DEBUG nova.network.neutron [req-c774b874-4e92-4d07-a56d-94c18dd78afe req-d0dd7296-246d-4e03-a112-4d5022617e1d service nova] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.219577] env[61867]: DEBUG nova.network.neutron [req-c774b874-4e92-4d07-a56d-94c18dd78afe req-d0dd7296-246d-4e03-a112-4d5022617e1d service nova] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.311856] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.412781] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c647358-841c-4d1c-b8c9-718fdd84fa53 tempest-ServerShowV254Test-648374040 tempest-ServerShowV254Test-648374040-project-member] Lock "37ed9e18-8dba-459d-bc67-e3ce3f9cdaea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.624s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.480629] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.727685] env[61867]: DEBUG oslo_concurrency.lockutils [req-c774b874-4e92-4d07-a56d-94c18dd78afe req-d0dd7296-246d-4e03-a112-4d5022617e1d service nova] Releasing lock "refresh_cache-5dab4fcf-1a5e-493a-842b-b652e99d67ba" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.728560] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Acquired lock "refresh_cache-5dab4fcf-1a5e-493a-842b-b652e99d67ba" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.729404] env[61867]: DEBUG nova.network.neutron [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 714.753015] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-125f76bb-517e-41c7-a136-8fa20af58e1c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.759703] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c4c4f87-5236-49b0-845b-911e03bbea79 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.798298] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69cf737a-df80-4f00-9fd7-0d42fd52ae3d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.805374] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760dbf4a-fdc4-40c1-856e-e7625cd14204 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.818499] env[61867]: DEBUG nova.compute.provider_tree [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.250665] env[61867]: DEBUG nova.network.neutron [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.311075] env[61867]: DEBUG nova.network.neutron [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.324371] env[61867]: DEBUG nova.scheduler.client.report [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 715.672981] env[61867]: DEBUG nova.compute.manager [req-e726e8fb-2cd7-478b-ba0f-f3c6a5ec472e req-6a6a37a2-d5a7-48ca-99d3-2f2a352fd211 service nova] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Received event network-vif-deleted-e2d64557-ec03-4212-bfcc-6e3d8cc9f0f0 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 715.814264] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Releasing lock "refresh_cache-5dab4fcf-1a5e-493a-842b-b652e99d67ba" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.815207] env[61867]: DEBUG nova.compute.manager [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 715.815207] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 715.815207] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9324288a-732a-43f8-b95a-88264fa0fb6e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.825937] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da21eff-bbae-40f1-b551-1cc920ac910b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.836212] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.960s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.836788] env[61867]: ERROR nova.compute.manager [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 19b41515-c7df-4409-9145-b398f19f6eec, please check neutron logs for more information. [ 715.836788] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] Traceback (most recent call last): [ 715.836788] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 715.836788] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] self.driver.spawn(context, instance, image_meta, [ 715.836788] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 715.836788] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] self._vmops.spawn(context, instance, image_meta, injected_files, [ 715.836788] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 715.836788] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] vm_ref = self.build_virtual_machine(instance, [ 715.836788] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 715.836788] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] vif_infos = vmwarevif.get_vif_info(self._session, [ 715.836788] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 715.837176] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] for vif in network_info: [ 715.837176] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 715.837176] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] return self._sync_wrapper(fn, *args, **kwargs) [ 715.837176] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 715.837176] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] self.wait() [ 715.837176] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 715.837176] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] self[:] = self._gt.wait() [ 715.837176] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 715.837176] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] return self._exit_event.wait() [ 715.837176] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 715.837176] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] current.throw(*self._exc) [ 715.837176] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 715.837176] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] result = function(*args, **kwargs) [ 715.837676] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 715.837676] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] return func(*args, **kwargs) [ 715.837676] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 715.837676] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] raise e [ 715.837676] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.837676] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] nwinfo = self.network_api.allocate_for_instance( [ 715.837676] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 715.837676] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] created_port_ids = self._update_ports_for_instance( [ 715.837676] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 715.837676] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] with excutils.save_and_reraise_exception(): [ 715.837676] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.837676] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] self.force_reraise() [ 715.837676] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.838101] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] raise self.value [ 715.838101] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 715.838101] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] updated_port = self._update_port( [ 715.838101] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.838101] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] _ensure_no_port_binding_failure(port) [ 715.838101] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.838101] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] raise exception.PortBindingFailed(port_id=port['id']) [ 715.838101] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] nova.exception.PortBindingFailed: Binding failed for port 19b41515-c7df-4409-9145-b398f19f6eec, please check neutron logs for more information. [ 715.838101] env[61867]: ERROR nova.compute.manager [instance: 774e7c38-426a-441d-8252-3daa424be558] [ 715.838101] env[61867]: DEBUG nova.compute.utils [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Binding failed for port 19b41515-c7df-4409-9145-b398f19f6eec, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 715.838749] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.846s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.840219] env[61867]: INFO nova.compute.claims [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 715.843417] env[61867]: DEBUG nova.compute.manager [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Build of instance 774e7c38-426a-441d-8252-3daa424be558 was re-scheduled: Binding failed for port 19b41515-c7df-4409-9145-b398f19f6eec, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 715.843810] env[61867]: DEBUG nova.compute.manager [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 715.844046] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Acquiring lock "refresh_cache-774e7c38-426a-441d-8252-3daa424be558" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.844199] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Acquired lock "refresh_cache-774e7c38-426a-441d-8252-3daa424be558" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.844353] env[61867]: DEBUG nova.network.neutron [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 715.854238] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5dab4fcf-1a5e-493a-842b-b652e99d67ba could not be found. [ 715.854620] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 715.854620] env[61867]: INFO nova.compute.manager [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Took 0.04 seconds to destroy the instance on the hypervisor. [ 715.854861] env[61867]: DEBUG oslo.service.loopingcall [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 715.855616] env[61867]: DEBUG nova.compute.manager [-] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 715.855732] env[61867]: DEBUG nova.network.neutron [-] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 715.874317] env[61867]: DEBUG nova.network.neutron [-] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.379185] env[61867]: DEBUG nova.network.neutron [-] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.423101] env[61867]: DEBUG nova.network.neutron [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.540489] env[61867]: DEBUG nova.network.neutron [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.884020] env[61867]: INFO nova.compute.manager [-] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Took 1.03 seconds to deallocate network for instance. [ 716.887682] env[61867]: DEBUG nova.compute.claims [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 716.887872] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.043304] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Releasing lock "refresh_cache-774e7c38-426a-441d-8252-3daa424be558" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.043542] env[61867]: DEBUG nova.compute.manager [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 717.043724] env[61867]: DEBUG nova.compute.manager [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 717.043887] env[61867]: DEBUG nova.network.neutron [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 717.064606] env[61867]: DEBUG nova.network.neutron [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.241158] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e5c80af-04f4-41b7-8840-a832a8e6c903 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.248521] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f3496c-d9c4-426c-a578-f3bcc69fa9f2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.280261] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fdb34cf-1021-4fdc-bc6a-e579e157cce8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.287447] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8551583c-644e-4f56-8f05-eb719c66d158 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.305849] env[61867]: DEBUG nova.compute.provider_tree [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.568469] env[61867]: DEBUG nova.network.neutron [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.811054] env[61867]: DEBUG nova.scheduler.client.report [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.072929] env[61867]: INFO nova.compute.manager [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] [instance: 774e7c38-426a-441d-8252-3daa424be558] Took 1.03 seconds to deallocate network for instance. [ 718.314917] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.476s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.315608] env[61867]: DEBUG nova.compute.manager [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 718.318977] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.153s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.766417] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Acquiring lock "4653ce96-257e-46d8-8c3d-85c03380213c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.766693] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Lock "4653ce96-257e-46d8-8c3d-85c03380213c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.823594] env[61867]: DEBUG nova.compute.utils [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 718.829303] env[61867]: DEBUG nova.compute.manager [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 718.829561] env[61867]: DEBUG nova.network.neutron [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 718.881612] env[61867]: DEBUG nova.policy [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1a29e34abfc4ce7873f7bf7d96eb474', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd177558d263c4c8295ef9406bd4aee99', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 719.118014] env[61867]: INFO nova.scheduler.client.report [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Deleted allocations for instance 774e7c38-426a-441d-8252-3daa424be558 [ 719.190255] env[61867]: DEBUG nova.network.neutron [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Successfully created port: d1b94eee-4eda-46c9-b144-f5356b51897e {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 719.195202] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d749c82-71a7-44d2-9189-35f19d5474f7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.203772] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a24d9671-3f9f-4410-b490-7d847e11a37b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.243090] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a13b63-4847-4fe6-a08b-632aa38a3ad5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.251412] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a9ce78-bb5d-420f-88e6-3b3d556392b7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.266750] env[61867]: DEBUG nova.compute.provider_tree [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.336210] env[61867]: DEBUG nova.compute.manager [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 719.626384] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c08f16b7-bcb2-4bbb-8000-7f2d58f3b7ba tempest-InstanceActionsTestJSON-311668165 tempest-InstanceActionsTestJSON-311668165-project-member] Lock "774e7c38-426a-441d-8252-3daa424be558" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.719s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.770271] env[61867]: DEBUG nova.scheduler.client.report [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.092217] env[61867]: DEBUG nova.compute.manager [req-e2712787-662f-45ea-8b4a-5fb0a3b36bdd req-e9889c39-9a5d-4204-8c88-68a6654aa090 service nova] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Received event network-changed-d1b94eee-4eda-46c9-b144-f5356b51897e {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.092423] env[61867]: DEBUG nova.compute.manager [req-e2712787-662f-45ea-8b4a-5fb0a3b36bdd req-e9889c39-9a5d-4204-8c88-68a6654aa090 service nova] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Refreshing instance network info cache due to event network-changed-d1b94eee-4eda-46c9-b144-f5356b51897e. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 720.092636] env[61867]: DEBUG oslo_concurrency.lockutils [req-e2712787-662f-45ea-8b4a-5fb0a3b36bdd req-e9889c39-9a5d-4204-8c88-68a6654aa090 service nova] Acquiring lock "refresh_cache-ab18091f-bab5-4926-b7a0-1f0e005bebe0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.092778] env[61867]: DEBUG oslo_concurrency.lockutils [req-e2712787-662f-45ea-8b4a-5fb0a3b36bdd req-e9889c39-9a5d-4204-8c88-68a6654aa090 service nova] Acquired lock "refresh_cache-ab18091f-bab5-4926-b7a0-1f0e005bebe0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.092932] env[61867]: DEBUG nova.network.neutron [req-e2712787-662f-45ea-8b4a-5fb0a3b36bdd req-e9889c39-9a5d-4204-8c88-68a6654aa090 service nova] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Refreshing network info cache for port d1b94eee-4eda-46c9-b144-f5356b51897e {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 720.128914] env[61867]: DEBUG nova.compute.manager [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 720.276873] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.957s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.276873] env[61867]: ERROR nova.compute.manager [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f0f2c586-6ec6-41cf-8afc-6e19ae0ae039, please check neutron logs for more information. [ 720.276873] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Traceback (most recent call last): [ 720.276873] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 720.276873] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] self.driver.spawn(context, instance, image_meta, [ 720.276873] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 720.276873] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.276873] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.276873] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] vm_ref = self.build_virtual_machine(instance, [ 720.277228] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.277228] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.277228] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.277228] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] for vif in network_info: [ 720.277228] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 720.277228] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] return self._sync_wrapper(fn, *args, **kwargs) [ 720.277228] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 720.277228] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] self.wait() [ 720.277228] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 720.277228] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] self[:] = self._gt.wait() [ 720.277228] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.277228] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] return self._exit_event.wait() [ 720.277228] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 720.277620] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] current.throw(*self._exc) [ 720.277620] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.277620] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] result = function(*args, **kwargs) [ 720.277620] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.277620] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] return func(*args, **kwargs) [ 720.277620] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.277620] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] raise e [ 720.277620] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.277620] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] nwinfo = self.network_api.allocate_for_instance( [ 720.277620] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.277620] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] created_port_ids = self._update_ports_for_instance( [ 720.277620] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.277620] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] with excutils.save_and_reraise_exception(): [ 720.278093] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.278093] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] self.force_reraise() [ 720.278093] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.278093] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] raise self.value [ 720.278093] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.278093] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] updated_port = self._update_port( [ 720.278093] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.278093] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] _ensure_no_port_binding_failure(port) [ 720.278093] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.278093] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] raise exception.PortBindingFailed(port_id=port['id']) [ 720.278093] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] nova.exception.PortBindingFailed: Binding failed for port f0f2c586-6ec6-41cf-8afc-6e19ae0ae039, please check neutron logs for more information. [ 720.278093] env[61867]: ERROR nova.compute.manager [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] [ 720.278790] env[61867]: DEBUG nova.compute.utils [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Binding failed for port f0f2c586-6ec6-41cf-8afc-6e19ae0ae039, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 720.278982] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.005s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.284575] env[61867]: DEBUG nova.compute.manager [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Build of instance 69d292d9-2330-47ac-94d4-6797abb8c167 was re-scheduled: Binding failed for port f0f2c586-6ec6-41cf-8afc-6e19ae0ae039, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 720.284575] env[61867]: DEBUG nova.compute.manager [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 720.284575] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Acquiring lock "refresh_cache-69d292d9-2330-47ac-94d4-6797abb8c167" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.284575] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Acquired lock "refresh_cache-69d292d9-2330-47ac-94d4-6797abb8c167" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.285019] env[61867]: DEBUG nova.network.neutron [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 720.344521] env[61867]: DEBUG nova.compute.manager [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 720.372130] env[61867]: DEBUG nova.virt.hardware [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 720.372224] env[61867]: DEBUG nova.virt.hardware [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 720.372300] env[61867]: DEBUG nova.virt.hardware [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 720.372472] env[61867]: DEBUG nova.virt.hardware [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 720.372611] env[61867]: DEBUG nova.virt.hardware [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 720.373054] env[61867]: DEBUG nova.virt.hardware [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 720.373054] env[61867]: DEBUG nova.virt.hardware [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 720.376124] env[61867]: DEBUG nova.virt.hardware [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 720.376124] env[61867]: DEBUG nova.virt.hardware [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 720.376124] env[61867]: DEBUG nova.virt.hardware [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 720.376124] env[61867]: DEBUG nova.virt.hardware [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.376124] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5424925e-a921-4e6b-b32a-5089beb4510e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.378906] env[61867]: ERROR nova.compute.manager [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d1b94eee-4eda-46c9-b144-f5356b51897e, please check neutron logs for more information. [ 720.378906] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 720.378906] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.378906] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 720.378906] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.378906] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 720.378906] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.378906] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 720.378906] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.378906] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 720.378906] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.378906] env[61867]: ERROR nova.compute.manager raise self.value [ 720.378906] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.378906] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 720.378906] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.378906] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 720.379544] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.379544] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 720.379544] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d1b94eee-4eda-46c9-b144-f5356b51897e, please check neutron logs for more information. [ 720.379544] env[61867]: ERROR nova.compute.manager [ 720.379544] env[61867]: Traceback (most recent call last): [ 720.379544] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 720.379544] env[61867]: listener.cb(fileno) [ 720.379544] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.379544] env[61867]: result = function(*args, **kwargs) [ 720.379544] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.379544] env[61867]: return func(*args, **kwargs) [ 720.379544] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.379544] env[61867]: raise e [ 720.379544] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.379544] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 720.379544] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.379544] env[61867]: created_port_ids = self._update_ports_for_instance( [ 720.379544] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.379544] env[61867]: with excutils.save_and_reraise_exception(): [ 720.379544] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.379544] env[61867]: self.force_reraise() [ 720.379544] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.379544] env[61867]: raise self.value [ 720.379544] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.379544] env[61867]: updated_port = self._update_port( [ 720.379544] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.379544] env[61867]: _ensure_no_port_binding_failure(port) [ 720.379544] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.379544] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 720.380477] env[61867]: nova.exception.PortBindingFailed: Binding failed for port d1b94eee-4eda-46c9-b144-f5356b51897e, please check neutron logs for more information. [ 720.380477] env[61867]: Removing descriptor: 21 [ 720.384599] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0040856e-5931-4061-a29b-b723383c7a2b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.398092] env[61867]: ERROR nova.compute.manager [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d1b94eee-4eda-46c9-b144-f5356b51897e, please check neutron logs for more information. [ 720.398092] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Traceback (most recent call last): [ 720.398092] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 720.398092] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] yield resources [ 720.398092] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 720.398092] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] self.driver.spawn(context, instance, image_meta, [ 720.398092] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 720.398092] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.398092] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.398092] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] vm_ref = self.build_virtual_machine(instance, [ 720.398092] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.398537] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.398537] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.398537] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] for vif in network_info: [ 720.398537] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 720.398537] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] return self._sync_wrapper(fn, *args, **kwargs) [ 720.398537] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 720.398537] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] self.wait() [ 720.398537] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 720.398537] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] self[:] = self._gt.wait() [ 720.398537] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.398537] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] return self._exit_event.wait() [ 720.398537] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 720.398537] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] current.throw(*self._exc) [ 720.399137] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.399137] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] result = function(*args, **kwargs) [ 720.399137] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.399137] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] return func(*args, **kwargs) [ 720.399137] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.399137] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] raise e [ 720.399137] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.399137] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] nwinfo = self.network_api.allocate_for_instance( [ 720.399137] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.399137] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] created_port_ids = self._update_ports_for_instance( [ 720.399137] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.399137] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] with excutils.save_and_reraise_exception(): [ 720.399137] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.399641] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] self.force_reraise() [ 720.399641] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.399641] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] raise self.value [ 720.399641] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.399641] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] updated_port = self._update_port( [ 720.399641] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.399641] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] _ensure_no_port_binding_failure(port) [ 720.399641] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.399641] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] raise exception.PortBindingFailed(port_id=port['id']) [ 720.399641] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] nova.exception.PortBindingFailed: Binding failed for port d1b94eee-4eda-46c9-b144-f5356b51897e, please check neutron logs for more information. [ 720.399641] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] [ 720.399641] env[61867]: INFO nova.compute.manager [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Terminating instance [ 720.400561] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "refresh_cache-ab18091f-bab5-4926-b7a0-1f0e005bebe0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.617016] env[61867]: DEBUG nova.network.neutron [req-e2712787-662f-45ea-8b4a-5fb0a3b36bdd req-e9889c39-9a5d-4204-8c88-68a6654aa090 service nova] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.654756] env[61867]: DEBUG oslo_concurrency.lockutils [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.698118] env[61867]: DEBUG nova.network.neutron [req-e2712787-662f-45ea-8b4a-5fb0a3b36bdd req-e9889c39-9a5d-4204-8c88-68a6654aa090 service nova] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.803552] env[61867]: DEBUG nova.network.neutron [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.884542] env[61867]: DEBUG nova.network.neutron [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.104280] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f17ada2-837a-4748-aaa9-b13b65ec10b6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.112788] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23038e49-9cc2-441e-8b0b-a199f06d4c53 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.145708] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ceb304b-77ba-43c0-a092-d2214cbde1a9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.153504] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08f268a7-c324-4293-89ef-6adc3ebc5aa9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.168678] env[61867]: DEBUG nova.compute.provider_tree [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.202019] env[61867]: DEBUG oslo_concurrency.lockutils [req-e2712787-662f-45ea-8b4a-5fb0a3b36bdd req-e9889c39-9a5d-4204-8c88-68a6654aa090 service nova] Releasing lock "refresh_cache-ab18091f-bab5-4926-b7a0-1f0e005bebe0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.202019] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquired lock "refresh_cache-ab18091f-bab5-4926-b7a0-1f0e005bebe0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.202019] env[61867]: DEBUG nova.network.neutron [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 721.389155] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Releasing lock "refresh_cache-69d292d9-2330-47ac-94d4-6797abb8c167" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.389487] env[61867]: DEBUG nova.compute.manager [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 721.389611] env[61867]: DEBUG nova.compute.manager [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 721.389773] env[61867]: DEBUG nova.network.neutron [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 721.405044] env[61867]: DEBUG nova.network.neutron [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.671850] env[61867]: DEBUG nova.scheduler.client.report [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 721.717580] env[61867]: DEBUG nova.network.neutron [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.800241] env[61867]: DEBUG nova.network.neutron [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.910825] env[61867]: DEBUG nova.network.neutron [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.135204] env[61867]: DEBUG nova.compute.manager [req-d1054451-9abf-485a-a7a4-f36e19b72394 req-c645c180-f458-4b58-9a78-52ca28d706e9 service nova] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Received event network-vif-deleted-d1b94eee-4eda-46c9-b144-f5356b51897e {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 722.177019] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.898s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.177662] env[61867]: ERROR nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0ccf49d0-6d5d-491b-a98a-d0afda9b598c, please check neutron logs for more information. [ 722.177662] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Traceback (most recent call last): [ 722.177662] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 722.177662] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] self.driver.spawn(context, instance, image_meta, [ 722.177662] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 722.177662] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 722.177662] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 722.177662] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] vm_ref = self.build_virtual_machine(instance, [ 722.177662] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 722.177662] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 722.177662] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 722.178107] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] for vif in network_info: [ 722.178107] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 722.178107] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] return self._sync_wrapper(fn, *args, **kwargs) [ 722.178107] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 722.178107] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] self.wait() [ 722.178107] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 722.178107] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] self[:] = self._gt.wait() [ 722.178107] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 722.178107] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] return self._exit_event.wait() [ 722.178107] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 722.178107] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] current.throw(*self._exc) [ 722.178107] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.178107] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] result = function(*args, **kwargs) [ 722.178546] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 722.178546] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] return func(*args, **kwargs) [ 722.178546] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.178546] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] raise e [ 722.178546] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.178546] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] nwinfo = self.network_api.allocate_for_instance( [ 722.178546] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.178546] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] created_port_ids = self._update_ports_for_instance( [ 722.178546] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.178546] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] with excutils.save_and_reraise_exception(): [ 722.178546] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.178546] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] self.force_reraise() [ 722.178546] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.179129] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] raise self.value [ 722.179129] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.179129] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] updated_port = self._update_port( [ 722.179129] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.179129] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] _ensure_no_port_binding_failure(port) [ 722.179129] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.179129] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] raise exception.PortBindingFailed(port_id=port['id']) [ 722.179129] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] nova.exception.PortBindingFailed: Binding failed for port 0ccf49d0-6d5d-491b-a98a-d0afda9b598c, please check neutron logs for more information. [ 722.179129] env[61867]: ERROR nova.compute.manager [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] [ 722.179129] env[61867]: DEBUG nova.compute.utils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Binding failed for port 0ccf49d0-6d5d-491b-a98a-d0afda9b598c, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 722.179582] env[61867]: DEBUG oslo_concurrency.lockutils [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.574s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.181058] env[61867]: INFO nova.compute.claims [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 722.184691] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Build of instance 3f1696b0-4de4-4128-bc83-e539e48dc8e3 was re-scheduled: Binding failed for port 0ccf49d0-6d5d-491b-a98a-d0afda9b598c, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 722.185130] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 722.185357] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquiring lock "refresh_cache-3f1696b0-4de4-4128-bc83-e539e48dc8e3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.185504] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquired lock "refresh_cache-3f1696b0-4de4-4128-bc83-e539e48dc8e3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.185662] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 722.302971] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Releasing lock "refresh_cache-ab18091f-bab5-4926-b7a0-1f0e005bebe0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.302971] env[61867]: DEBUG nova.compute.manager [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 722.303114] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 722.303367] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a3d678a8-ac9e-4bc0-bd0d-1216201523ab {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.312320] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e616406-96b7-4b28-abfa-37169588e852 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.333835] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ab18091f-bab5-4926-b7a0-1f0e005bebe0 could not be found. [ 722.335029] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 722.335029] env[61867]: INFO nova.compute.manager [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Took 0.03 seconds to destroy the instance on the hypervisor. [ 722.335029] env[61867]: DEBUG oslo.service.loopingcall [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.335029] env[61867]: DEBUG nova.compute.manager [-] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.335029] env[61867]: DEBUG nova.network.neutron [-] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 722.348727] env[61867]: DEBUG nova.network.neutron [-] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.413820] env[61867]: INFO nova.compute.manager [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] [instance: 69d292d9-2330-47ac-94d4-6797abb8c167] Took 1.02 seconds to deallocate network for instance. [ 722.706348] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.784717] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.850950] env[61867]: DEBUG nova.network.neutron [-] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.289914] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Releasing lock "refresh_cache-3f1696b0-4de4-4128-bc83-e539e48dc8e3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.290266] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 723.290391] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 723.290555] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 723.328237] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.357078] env[61867]: INFO nova.compute.manager [-] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Took 1.02 seconds to deallocate network for instance. [ 723.361138] env[61867]: DEBUG nova.compute.claims [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 723.361331] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.463889] env[61867]: INFO nova.scheduler.client.report [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Deleted allocations for instance 69d292d9-2330-47ac-94d4-6797abb8c167 [ 723.591031] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1baa40ed-fbd1-4c7d-abe8-92ccaae8e503 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.599458] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a91be00-dff8-4200-a4af-acce41dcaf80 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.630597] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d0e5dbf-5e99-4827-bd33-87518aefcd3e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.638486] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb1c7f5a-6bf2-4f07-95cd-7e782c74b5ec {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.652637] env[61867]: DEBUG nova.compute.provider_tree [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.831540] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.982067] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f62cb5c-a752-4ac7-a716-dc812f078190 tempest-ServersTestBootFromVolume-270901222 tempest-ServersTestBootFromVolume-270901222-project-member] Lock "69d292d9-2330-47ac-94d4-6797abb8c167" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.303s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.155656] env[61867]: DEBUG nova.scheduler.client.report [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 724.334469] env[61867]: INFO nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 3f1696b0-4de4-4128-bc83-e539e48dc8e3] Took 1.04 seconds to deallocate network for instance. [ 724.484955] env[61867]: DEBUG nova.compute.manager [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 724.663156] env[61867]: DEBUG oslo_concurrency.lockutils [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.663702] env[61867]: DEBUG nova.compute.manager [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 724.667177] env[61867]: DEBUG oslo_concurrency.lockutils [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.167s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.669924] env[61867]: INFO nova.compute.claims [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 725.012211] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.177538] env[61867]: DEBUG nova.compute.utils [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 725.181062] env[61867]: DEBUG nova.compute.manager [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Not allocating networking since 'none' was specified. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 725.209812] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Acquiring lock "66518b2a-0242-438b-ba9f-d57c07a1165c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.210412] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Lock "66518b2a-0242-438b-ba9f-d57c07a1165c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.378019] env[61867]: INFO nova.scheduler.client.report [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Deleted allocations for instance 3f1696b0-4de4-4128-bc83-e539e48dc8e3 [ 725.682743] env[61867]: DEBUG nova.compute.manager [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 725.888945] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "3f1696b0-4de4-4128-bc83-e539e48dc8e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.500s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.058946] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84717358-195b-4d60-a3b1-c0c2c11be43f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.069275] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa5de1c0-35c7-46fd-84a9-d1c9b0dda93e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.112506] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-181a3e5c-dc56-41f9-94d4-546732fc9c8b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.121857] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1846dff-3847-4970-9905-023ae731bdeb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.144345] env[61867]: DEBUG nova.compute.provider_tree [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.396546] env[61867]: DEBUG nova.compute.manager [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 726.650077] env[61867]: DEBUG nova.scheduler.client.report [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.698302] env[61867]: DEBUG nova.compute.manager [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 726.727296] env[61867]: DEBUG nova.virt.hardware [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 726.727559] env[61867]: DEBUG nova.virt.hardware [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 726.727746] env[61867]: DEBUG nova.virt.hardware [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 726.727881] env[61867]: DEBUG nova.virt.hardware [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 726.728032] env[61867]: DEBUG nova.virt.hardware [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 726.728181] env[61867]: DEBUG nova.virt.hardware [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 726.728387] env[61867]: DEBUG nova.virt.hardware [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 726.728542] env[61867]: DEBUG nova.virt.hardware [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 726.729063] env[61867]: DEBUG nova.virt.hardware [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 726.729063] env[61867]: DEBUG nova.virt.hardware [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 726.729063] env[61867]: DEBUG nova.virt.hardware [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 726.729936] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac1cb79f-a2e1-4403-93cc-c34afd6e35f0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.738720] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9d5efe4-f522-40b1-8745-8200a62d0ede {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.752811] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Instance VIF info [] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 726.758124] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Creating folder: Project (173a90df2be440baa325188dd0856907). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 726.759565] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5372f0fc-baac-46d2-8dbb-6f9eaa1d87d1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.770268] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Created folder: Project (173a90df2be440baa325188dd0856907) in parent group-v274258. [ 726.770519] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Creating folder: Instances. Parent ref: group-v274279. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 726.770768] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e7767301-c8be-4381-bb90-9e7a691e6a7d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.779530] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Created folder: Instances in parent group-v274279. [ 726.779781] env[61867]: DEBUG oslo.service.loopingcall [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 726.779967] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 726.780203] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f1746882-f16a-494b-9f77-ec25ece8f080 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.799016] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 726.799016] env[61867]: value = "task-1276368" [ 726.799016] env[61867]: _type = "Task" [ 726.799016] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.807840] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276368, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.930995] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.156402] env[61867]: DEBUG oslo_concurrency.lockutils [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.489s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.157105] env[61867]: DEBUG nova.compute.manager [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 727.159630] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.980s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.309797] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276368, 'name': CreateVM_Task, 'duration_secs': 0.266125} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.309960] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 727.310499] env[61867]: DEBUG oslo_concurrency.lockutils [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.310637] env[61867]: DEBUG oslo_concurrency.lockutils [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.310988] env[61867]: DEBUG oslo_concurrency.lockutils [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 727.311253] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78b689e0-7269-4549-a05b-5e814e636faf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.315830] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 727.315830] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526a60df-10bb-de6c-9ca0-5581e00e1a3e" [ 727.315830] env[61867]: _type = "Task" [ 727.315830] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.324092] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526a60df-10bb-de6c-9ca0-5581e00e1a3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.667474] env[61867]: DEBUG nova.compute.utils [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 727.668991] env[61867]: DEBUG nova.compute.manager [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Not allocating networking since 'none' was specified. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 727.826133] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526a60df-10bb-de6c-9ca0-5581e00e1a3e, 'name': SearchDatastore_Task, 'duration_secs': 0.009207} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.826429] env[61867]: DEBUG oslo_concurrency.lockutils [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.826656] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 727.826878] env[61867]: DEBUG oslo_concurrency.lockutils [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.827031] env[61867]: DEBUG oslo_concurrency.lockutils [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.827251] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 727.827514] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-50d882c8-1491-410b-90f5-29e1a91df5c0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.835756] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 727.835756] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 727.838071] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c4a64ff-75a0-4e03-a9c5-c29326f96517 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.843098] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 727.843098] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ca0f18-3b02-b95b-fd8a-53fff8f4b054" [ 727.843098] env[61867]: _type = "Task" [ 727.843098] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.850565] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ca0f18-3b02-b95b-fd8a-53fff8f4b054, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.995167] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc51ab75-0bf1-41d9-8dfd-625ba7e494e1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.003100] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b2ab56-f334-47a5-a6c3-23eb6c934e51 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.034855] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0084c833-8b90-4a9f-ac4c-169aaaeb97f4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.042613] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e24529-be66-4bfe-a1db-6d0dd07c4db0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.056818] env[61867]: DEBUG nova.compute.provider_tree [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.170714] env[61867]: DEBUG nova.compute.manager [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 728.353722] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ca0f18-3b02-b95b-fd8a-53fff8f4b054, 'name': SearchDatastore_Task, 'duration_secs': 0.007778} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.354537] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2bae781-7096-410f-bbc2-3a920ff01a2b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.359498] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 728.359498] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523bf880-93aa-0911-a544-8ed8c973b1b0" [ 728.359498] env[61867]: _type = "Task" [ 728.359498] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.366839] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523bf880-93aa-0911-a544-8ed8c973b1b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.559951] env[61867]: DEBUG nova.scheduler.client.report [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 728.871049] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523bf880-93aa-0911-a544-8ed8c973b1b0, 'name': SearchDatastore_Task, 'duration_secs': 0.008884} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.871694] env[61867]: DEBUG oslo_concurrency.lockutils [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.872085] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] b1ec32c5-0642-4dca-ad43-05a5172d04d7/b1ec32c5-0642-4dca-ad43-05a5172d04d7.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 728.872503] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-70b5f77c-3298-4a6a-b2ba-bc669224d008 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.878768] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 728.878768] env[61867]: value = "task-1276369" [ 728.878768] env[61867]: _type = "Task" [ 728.878768] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.888206] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276369, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.065074] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.905s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.065682] env[61867]: ERROR nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fddc7c36-1b1e-4671-a7dd-fe475981f035, please check neutron logs for more information. [ 729.065682] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Traceback (most recent call last): [ 729.065682] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 729.065682] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] self.driver.spawn(context, instance, image_meta, [ 729.065682] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 729.065682] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] self._vmops.spawn(context, instance, image_meta, injected_files, [ 729.065682] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 729.065682] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] vm_ref = self.build_virtual_machine(instance, [ 729.065682] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 729.065682] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] vif_infos = vmwarevif.get_vif_info(self._session, [ 729.065682] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 729.066122] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] for vif in network_info: [ 729.066122] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 729.066122] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] return self._sync_wrapper(fn, *args, **kwargs) [ 729.066122] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 729.066122] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] self.wait() [ 729.066122] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 729.066122] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] self[:] = self._gt.wait() [ 729.066122] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 729.066122] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] return self._exit_event.wait() [ 729.066122] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 729.066122] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] current.throw(*self._exc) [ 729.066122] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.066122] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] result = function(*args, **kwargs) [ 729.066601] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 729.066601] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] return func(*args, **kwargs) [ 729.066601] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.066601] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] raise e [ 729.066601] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.066601] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] nwinfo = self.network_api.allocate_for_instance( [ 729.066601] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.066601] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] created_port_ids = self._update_ports_for_instance( [ 729.066601] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.066601] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] with excutils.save_and_reraise_exception(): [ 729.066601] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.066601] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] self.force_reraise() [ 729.066601] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.067123] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] raise self.value [ 729.067123] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.067123] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] updated_port = self._update_port( [ 729.067123] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.067123] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] _ensure_no_port_binding_failure(port) [ 729.067123] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.067123] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] raise exception.PortBindingFailed(port_id=port['id']) [ 729.067123] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] nova.exception.PortBindingFailed: Binding failed for port fddc7c36-1b1e-4671-a7dd-fe475981f035, please check neutron logs for more information. [ 729.067123] env[61867]: ERROR nova.compute.manager [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] [ 729.067123] env[61867]: DEBUG nova.compute.utils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Binding failed for port fddc7c36-1b1e-4671-a7dd-fe475981f035, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 729.068596] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.124s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.072056] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Build of instance 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de was re-scheduled: Binding failed for port fddc7c36-1b1e-4671-a7dd-fe475981f035, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 729.072605] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 729.072940] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquiring lock "refresh_cache-2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.074130] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquired lock "refresh_cache-2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.074130] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 729.180958] env[61867]: DEBUG nova.compute.manager [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 729.212654] env[61867]: DEBUG nova.virt.hardware [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 729.212935] env[61867]: DEBUG nova.virt.hardware [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 729.213356] env[61867]: DEBUG nova.virt.hardware [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 729.213356] env[61867]: DEBUG nova.virt.hardware [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 729.213485] env[61867]: DEBUG nova.virt.hardware [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 729.213549] env[61867]: DEBUG nova.virt.hardware [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 729.213746] env[61867]: DEBUG nova.virt.hardware [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 729.213907] env[61867]: DEBUG nova.virt.hardware [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 729.214370] env[61867]: DEBUG nova.virt.hardware [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 729.214370] env[61867]: DEBUG nova.virt.hardware [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 729.214567] env[61867]: DEBUG nova.virt.hardware [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 729.215314] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852fb4e8-7083-43a0-9f1d-1805dfca686a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.228576] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16404f06-3d40-48ed-9065-0db922ebd41d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.245495] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Instance VIF info [] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 729.251795] env[61867]: DEBUG oslo.service.loopingcall [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 729.252780] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 729.252891] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b4f477af-9db6-4286-aedc-f8dacb1a5697 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.271967] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 729.271967] env[61867]: value = "task-1276370" [ 729.271967] env[61867]: _type = "Task" [ 729.271967] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.284115] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276370, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.396961] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276369, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446801} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.397297] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] b1ec32c5-0642-4dca-ad43-05a5172d04d7/b1ec32c5-0642-4dca-ad43-05a5172d04d7.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 729.397546] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 729.397774] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e40ca987-1232-4619-b880-1891c9cfc933 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.404480] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 729.404480] env[61867]: value = "task-1276371" [ 729.404480] env[61867]: _type = "Task" [ 729.404480] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.414741] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276371, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.610154] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.703031] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.782561] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276370, 'name': CreateVM_Task, 'duration_secs': 0.288937} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.782734] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 729.783153] env[61867]: DEBUG oslo_concurrency.lockutils [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.783359] env[61867]: DEBUG oslo_concurrency.lockutils [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.783669] env[61867]: DEBUG oslo_concurrency.lockutils [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 729.783910] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c95e630-6898-4f10-b5c6-fc15193f392a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.788451] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 729.788451] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52524c9d-bebe-b864-5dc8-87b89f4a6f04" [ 729.788451] env[61867]: _type = "Task" [ 729.788451] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.796062] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52524c9d-bebe-b864-5dc8-87b89f4a6f04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.914778] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276371, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072717} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.915033] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 729.915908] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae7261d7-17d8-41c3-a4d1-ea2ec526d622 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.920306] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba49bb6-2d27-4ba3-845e-1c0f0de25c7e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.936659] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99cc85ae-6fb5-4775-9e81-ffe0bd673fbc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.946086] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] b1ec32c5-0642-4dca-ad43-05a5172d04d7/b1ec32c5-0642-4dca-ad43-05a5172d04d7.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 729.946385] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ddb1fd26-32f5-41ca-a0c1-4c8e746885af {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.992634] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ce7c23-275f-40eb-8d05-12da86ea2e01 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.995647] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 729.995647] env[61867]: value = "task-1276372" [ 729.995647] env[61867]: _type = "Task" [ 729.995647] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.002625] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7fa4789-df8e-4c1d-a384-4dc1fd4220f1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.010375] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276372, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.019898] env[61867]: DEBUG nova.compute.provider_tree [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.204933] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Releasing lock "refresh_cache-2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.205210] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 730.205392] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 730.205877] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 730.225212] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.299711] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52524c9d-bebe-b864-5dc8-87b89f4a6f04, 'name': SearchDatastore_Task, 'duration_secs': 0.009179} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.300030] env[61867]: DEBUG oslo_concurrency.lockutils [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.300351] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 730.300579] env[61867]: DEBUG oslo_concurrency.lockutils [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.301605] env[61867]: DEBUG oslo_concurrency.lockutils [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.301605] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 730.301605] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6f671c5a-da21-4e60-ba81-1dc1c686aaa4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.309543] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 730.309543] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 730.309693] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed5b9a86-8e81-4a68-b75c-d48f8cfe35e2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.315053] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 730.315053] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f21ad3-cd28-bd60-6a1a-f113d4a441bd" [ 730.315053] env[61867]: _type = "Task" [ 730.315053] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.322690] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f21ad3-cd28-bd60-6a1a-f113d4a441bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.505470] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276372, 'name': ReconfigVM_Task, 'duration_secs': 0.267592} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.505768] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Reconfigured VM instance instance-0000002e to attach disk [datastore2] b1ec32c5-0642-4dca-ad43-05a5172d04d7/b1ec32c5-0642-4dca-ad43-05a5172d04d7.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 730.506435] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-15d1e427-93e7-444e-a7b6-e23326cd5b76 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.512151] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 730.512151] env[61867]: value = "task-1276373" [ 730.512151] env[61867]: _type = "Task" [ 730.512151] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.519760] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276373, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.523059] env[61867]: DEBUG nova.scheduler.client.report [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.729100] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.825807] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f21ad3-cd28-bd60-6a1a-f113d4a441bd, 'name': SearchDatastore_Task, 'duration_secs': 0.008611} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.826604] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc516f80-08e9-46d6-b927-2932fbb38bb4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.831733] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 730.831733] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520394fa-d0cf-0295-ca2a-6e25d604bd37" [ 730.831733] env[61867]: _type = "Task" [ 730.831733] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.840219] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520394fa-d0cf-0295-ca2a-6e25d604bd37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.022389] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276373, 'name': Rename_Task, 'duration_secs': 0.140446} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.022678] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 731.022905] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9a7b3584-2718-45b2-aab8-ce8be1c4485a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.027255] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.959s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.027885] env[61867]: ERROR nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b13b1e49-c2b5-4a52-adec-93dda37f48af, please check neutron logs for more information. [ 731.027885] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Traceback (most recent call last): [ 731.027885] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 731.027885] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] self.driver.spawn(context, instance, image_meta, [ 731.027885] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 731.027885] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 731.027885] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 731.027885] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] vm_ref = self.build_virtual_machine(instance, [ 731.027885] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 731.027885] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] vif_infos = vmwarevif.get_vif_info(self._session, [ 731.027885] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 731.028276] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] for vif in network_info: [ 731.028276] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 731.028276] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] return self._sync_wrapper(fn, *args, **kwargs) [ 731.028276] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 731.028276] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] self.wait() [ 731.028276] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 731.028276] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] self[:] = self._gt.wait() [ 731.028276] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 731.028276] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] return self._exit_event.wait() [ 731.028276] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 731.028276] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] current.throw(*self._exc) [ 731.028276] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.028276] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] result = function(*args, **kwargs) [ 731.028708] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 731.028708] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] return func(*args, **kwargs) [ 731.028708] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 731.028708] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] raise e [ 731.028708] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.028708] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] nwinfo = self.network_api.allocate_for_instance( [ 731.028708] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 731.028708] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] created_port_ids = self._update_ports_for_instance( [ 731.028708] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 731.028708] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] with excutils.save_and_reraise_exception(): [ 731.028708] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.028708] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] self.force_reraise() [ 731.028708] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.029267] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] raise self.value [ 731.029267] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 731.029267] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] updated_port = self._update_port( [ 731.029267] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.029267] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] _ensure_no_port_binding_failure(port) [ 731.029267] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.029267] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] raise exception.PortBindingFailed(port_id=port['id']) [ 731.029267] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] nova.exception.PortBindingFailed: Binding failed for port b13b1e49-c2b5-4a52-adec-93dda37f48af, please check neutron logs for more information. [ 731.029267] env[61867]: ERROR nova.compute.manager [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] [ 731.029267] env[61867]: DEBUG nova.compute.utils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Binding failed for port b13b1e49-c2b5-4a52-adec-93dda37f48af, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 731.030723] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.719s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.032191] env[61867]: INFO nova.compute.claims [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.035006] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 731.035006] env[61867]: value = "task-1276374" [ 731.035006] env[61867]: _type = "Task" [ 731.035006] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.035006] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Build of instance add3bbbb-fd30-49ad-b95e-601af5d790e4 was re-scheduled: Binding failed for port b13b1e49-c2b5-4a52-adec-93dda37f48af, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 731.035397] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 731.035617] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquiring lock "refresh_cache-add3bbbb-fd30-49ad-b95e-601af5d790e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.035765] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Acquired lock "refresh_cache-add3bbbb-fd30-49ad-b95e-601af5d790e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.036180] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 731.045240] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276374, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.122533] env[61867]: DEBUG oslo_concurrency.lockutils [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.122759] env[61867]: DEBUG oslo_concurrency.lockutils [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.231161] env[61867]: INFO nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de] Took 1.02 seconds to deallocate network for instance. [ 731.342046] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520394fa-d0cf-0295-ca2a-6e25d604bd37, 'name': SearchDatastore_Task, 'duration_secs': 0.008692} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.342382] env[61867]: DEBUG oslo_concurrency.lockutils [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.342649] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 8656c3b8-b7b1-442a-9719-1588ee83d19b/8656c3b8-b7b1-442a-9719-1588ee83d19b.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 731.342901] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc5cc3be-d08b-491e-8f8c-49dd02083b16 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.349065] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 731.349065] env[61867]: value = "task-1276375" [ 731.349065] env[61867]: _type = "Task" [ 731.349065] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.357665] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276375, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.552911] env[61867]: DEBUG oslo_vmware.api [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276374, 'name': PowerOnVM_Task, 'duration_secs': 0.410694} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.552911] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 731.552911] env[61867]: INFO nova.compute.manager [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Took 4.85 seconds to spawn the instance on the hypervisor. [ 731.553210] env[61867]: DEBUG nova.compute.manager [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 731.554027] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3225ca8-3a14-4b66-b4f1-4b3570fa8d82 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.563937] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.660715] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.858527] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276375, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.075780] env[61867]: INFO nova.compute.manager [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Took 27.49 seconds to build instance. [ 732.164375] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Releasing lock "refresh_cache-add3bbbb-fd30-49ad-b95e-601af5d790e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.164624] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 732.164801] env[61867]: DEBUG nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 732.164959] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 732.181935] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.262323] env[61867]: INFO nova.scheduler.client.report [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Deleted allocations for instance 2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de [ 732.360043] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276375, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.588645} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.362660] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 8656c3b8-b7b1-442a-9719-1588ee83d19b/8656c3b8-b7b1-442a-9719-1588ee83d19b.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 732.362842] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 732.363219] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3bc7f367-9d0f-44cf-9ba4-5b16d5bb7daa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.369344] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 732.369344] env[61867]: value = "task-1276376" [ 732.369344] env[61867]: _type = "Task" [ 732.369344] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.373699] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e944c35-46f4-4f6d-af9c-0f3f6c2467f7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.380702] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276376, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.383204] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8ab5da-4a93-4afb-a4d2-76cddd10cfde {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.412791] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8158249a-cd52-4804-b169-9c1b3124c2d7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.420535] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7318bd0b-900c-4127-aa6d-ac7d475af550 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.433211] env[61867]: DEBUG nova.compute.provider_tree [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.577730] env[61867]: DEBUG oslo_concurrency.lockutils [None req-81a1df6e-3e81-404a-af4f-4f23fbd47689 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "b1ec32c5-0642-4dca-ad43-05a5172d04d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.185s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.684380] env[61867]: DEBUG nova.network.neutron [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.772593] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "2b9b62b3-ebe1-4fe9-a63b-1003fca3d5de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.356s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.879892] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276376, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065005} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.883712] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 732.883712] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d8a66b-4e9d-4966-9a6b-ae3b9b2b117c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.900237] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] 8656c3b8-b7b1-442a-9719-1588ee83d19b/8656c3b8-b7b1-442a-9719-1588ee83d19b.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 732.903462] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aadec120-06d0-4c4c-84fc-b0f673c37c25 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.921256] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 732.921256] env[61867]: value = "task-1276377" [ 732.921256] env[61867]: _type = "Task" [ 732.921256] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.928850] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276377, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.935753] env[61867]: DEBUG nova.scheduler.client.report [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 733.079833] env[61867]: DEBUG nova.compute.manager [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 733.187267] env[61867]: INFO nova.compute.manager [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] [instance: add3bbbb-fd30-49ad-b95e-601af5d790e4] Took 1.02 seconds to deallocate network for instance. [ 733.277089] env[61867]: DEBUG nova.compute.manager [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 733.429768] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276377, 'name': ReconfigVM_Task, 'duration_secs': 0.269244} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.430105] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Reconfigured VM instance instance-0000002f to attach disk [datastore2] 8656c3b8-b7b1-442a-9719-1588ee83d19b/8656c3b8-b7b1-442a-9719-1588ee83d19b.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 733.430753] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cdf7bb52-d7aa-4597-b2cf-c6c47342476a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.437545] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 733.437545] env[61867]: value = "task-1276378" [ 733.437545] env[61867]: _type = "Task" [ 733.437545] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.441371] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.441831] env[61867]: DEBUG nova.compute.manager [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 733.444445] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.964s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.445242] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.445420] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61867) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 733.445684] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.558s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.449730] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08956a9c-a7bf-481b-a2a5-3887489d96f1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.455024] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276378, 'name': Rename_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.459732] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed86d762-d819-4c72-a582-c60615e33f5c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.473457] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcea724e-0997-4d75-a4be-54ee0f9136a8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.480582] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbf994e9-31da-41ce-937f-c9b50a058a1c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.510135] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181518MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61867) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 733.510323] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.603685] env[61867]: DEBUG oslo_concurrency.lockutils [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.798239] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.948553] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276378, 'name': Rename_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.950284] env[61867]: DEBUG nova.compute.utils [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 733.954881] env[61867]: DEBUG nova.compute.manager [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 733.954881] env[61867]: DEBUG nova.network.neutron [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 734.052456] env[61867]: DEBUG nova.policy [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cebbcc4f2c2c47fe8f4aae84550cb17c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8423840c6aec4fc4a6c6887d0f844b19', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 734.314433] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72cbd133-3874-4631-90db-54a434481adb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.321983] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8117016-99fc-40ce-9420-db0ca05ec1ac {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.354289] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f7a07a-3ec3-4ee0-a900-8be967d43671 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.362207] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8cb691-ad82-4f58-a9e9-211a0b8351d6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.376260] env[61867]: DEBUG nova.compute.provider_tree [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.447558] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276378, 'name': Rename_Task, 'duration_secs': 0.829648} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.447826] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 734.448076] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-db42c304-0cd8-4cbc-bdf8-f81c2c65e8e5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.455494] env[61867]: DEBUG nova.compute.manager [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 734.462019] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 734.462019] env[61867]: value = "task-1276379" [ 734.462019] env[61867]: _type = "Task" [ 734.462019] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.468346] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276379, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.474206] env[61867]: DEBUG nova.network.neutron [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Successfully created port: bfc894ba-9915-48bc-b7b2-83785909f7d4 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 734.879246] env[61867]: DEBUG nova.scheduler.client.report [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.975373] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276379, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.230194] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d1fc8bd9-6f8e-4228-a979-fe330a7e255a tempest-ListServersNegativeTestJSON-1774581571 tempest-ListServersNegativeTestJSON-1774581571-project-member] Lock "add3bbbb-fd30-49ad-b95e-601af5d790e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.784s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.231104] env[61867]: Traceback (most recent call last): [ 735.231104] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 735.231104] env[61867]: self.driver.spawn(context, instance, image_meta, [ 735.231104] env[61867]: File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 735.231104] env[61867]: self._vmops.spawn(context, instance, image_meta, injected_files, [ 735.231104] env[61867]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 735.231104] env[61867]: vm_ref = self.build_virtual_machine(instance, [ 735.231104] env[61867]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 735.231104] env[61867]: vif_infos = vmwarevif.get_vif_info(self._session, [ 735.231104] env[61867]: File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 735.231104] env[61867]: for vif in network_info: [ 735.231104] env[61867]: File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 735.231104] env[61867]: return self._sync_wrapper(fn, *args, **kwargs) [ 735.231104] env[61867]: File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 735.231104] env[61867]: self.wait() [ 735.231104] env[61867]: File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 735.231104] env[61867]: self[:] = self._gt.wait() [ 735.231104] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 735.231104] env[61867]: return self._exit_event.wait() [ 735.231104] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 735.231104] env[61867]: current.throw(*self._exc) [ 735.231104] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.231104] env[61867]: result = function(*args, **kwargs) [ 735.231104] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.231104] env[61867]: return func(*args, **kwargs) [ 735.231104] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.231104] env[61867]: raise e [ 735.231104] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.231104] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 735.231104] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.231104] env[61867]: created_port_ids = self._update_ports_for_instance( [ 735.231104] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.231104] env[61867]: with excutils.save_and_reraise_exception(): [ 735.233082] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.233082] env[61867]: self.force_reraise() [ 735.233082] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.233082] env[61867]: raise self.value [ 735.233082] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.233082] env[61867]: updated_port = self._update_port( [ 735.233082] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.233082] env[61867]: _ensure_no_port_binding_failure(port) [ 735.233082] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.233082] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 735.233082] env[61867]: nova.exception.PortBindingFailed: Binding failed for port b13b1e49-c2b5-4a52-adec-93dda37f48af, please check neutron logs for more information. [ 735.233082] env[61867]: During handling of the above exception, another exception occurred: [ 735.233082] env[61867]: Traceback (most recent call last): [ 735.233082] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2456, in _do_build_and_run_instance [ 735.233082] env[61867]: self._build_and_run_instance(context, instance, image, [ 735.233082] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2748, in _build_and_run_instance [ 735.233082] env[61867]: raise exception.RescheduledException( [ 735.233082] env[61867]: nova.exception.RescheduledException: Build of instance add3bbbb-fd30-49ad-b95e-601af5d790e4 was re-scheduled: Binding failed for port b13b1e49-c2b5-4a52-adec-93dda37f48af, please check neutron logs for more information. [ 735.233082] env[61867]: During handling of the above exception, another exception occurred: [ 735.233082] env[61867]: Traceback (most recent call last): [ 735.233082] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 735.233082] env[61867]: func(*args, **kwargs) [ 735.233082] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.233082] env[61867]: return func(*args, **kwargs) [ 735.233082] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 735.233082] env[61867]: return f(*args, **kwargs) [ 735.233082] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2347, in _locked_do_build_and_run_instance [ 735.233082] env[61867]: result = self._do_build_and_run_instance(*args, **kwargs) [ 735.235322] env[61867]: File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 735.235322] env[61867]: with excutils.save_and_reraise_exception(): [ 735.235322] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.235322] env[61867]: self.force_reraise() [ 735.235322] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.235322] env[61867]: raise self.value [ 735.235322] env[61867]: File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 735.235322] env[61867]: return f(self, context, *args, **kw) [ 735.235322] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 735.235322] env[61867]: with excutils.save_and_reraise_exception(): [ 735.235322] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.235322] env[61867]: self.force_reraise() [ 735.235322] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.235322] env[61867]: raise self.value [ 735.235322] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 735.235322] env[61867]: return function(self, context, *args, **kwargs) [ 735.235322] env[61867]: File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 735.235322] env[61867]: return function(self, context, *args, **kwargs) [ 735.235322] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 735.235322] env[61867]: return function(self, context, *args, **kwargs) [ 735.235322] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2491, in _do_build_and_run_instance [ 735.235322] env[61867]: instance.save() [ 735.235322] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 209, in wrapper [ 735.235322] env[61867]: updates, result = self.indirection_api.object_action( [ 735.235322] env[61867]: File "/opt/stack/nova/nova/conductor/rpcapi.py", line 247, in object_action [ 735.235322] env[61867]: return cctxt.call(context, 'object_action', objinst=objinst, [ 735.235322] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/client.py", line 190, in call [ 735.235322] env[61867]: result = self.transport._send( [ 735.235322] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/transport.py", line 123, in _send [ 735.235322] env[61867]: return self._driver.send(target, ctxt, message, [ 735.236308] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 788, in send [ 735.236308] env[61867]: return self._send(target, ctxt, message, wait_for_reply, timeout, [ 735.236308] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 780, in _send [ 735.236308] env[61867]: raise result [ 735.236308] env[61867]: nova.exception_Remote.InstanceNotFound_Remote: Instance add3bbbb-fd30-49ad-b95e-601af5d790e4 could not be found. [ 735.236308] env[61867]: Traceback (most recent call last): [ 735.236308] env[61867]: File "/opt/stack/nova/nova/conductor/manager.py", line 142, in _object_dispatch [ 735.236308] env[61867]: return getattr(target, method)(*args, **kwargs) [ 735.236308] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 226, in wrapper [ 735.236308] env[61867]: return fn(self, *args, **kwargs) [ 735.236308] env[61867]: File "/opt/stack/nova/nova/objects/instance.py", line 878, in save [ 735.236308] env[61867]: old_ref, inst_ref = db.instance_update_and_get_original( [ 735.236308] env[61867]: File "/opt/stack/nova/nova/db/utils.py", line 35, in wrapper [ 735.236308] env[61867]: return f(*args, **kwargs) [ 735.236308] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 144, in wrapper [ 735.236308] env[61867]: with excutils.save_and_reraise_exception() as ectxt: [ 735.236308] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.236308] env[61867]: self.force_reraise() [ 735.236308] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.236308] env[61867]: raise self.value [ 735.236308] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 142, in wrapper [ 735.236308] env[61867]: return f(*args, **kwargs) [ 735.236308] env[61867]: File "/opt/stack/nova/nova/db/main/api.py", line 207, in wrapper [ 735.236308] env[61867]: return f(context, *args, **kwargs) [ 735.236308] env[61867]: File "/opt/stack/nova/nova/db/main/api.py", line 2283, in instance_update_and_get_original [ 735.236308] env[61867]: instance_ref = _instance_get_by_uuid(context, instance_uuid, [ 735.236308] env[61867]: File "/opt/stack/nova/nova/db/main/api.py", line 1405, in _instance_get_by_uuid [ 735.236308] env[61867]: raise exception.InstanceNotFound(instance_id=uuid) [ 735.236308] env[61867]: nova.exception.InstanceNotFound: Instance add3bbbb-fd30-49ad-b95e-601af5d790e4 could not be found. [ 735.247833] env[61867]: DEBUG nova.compute.manager [req-b93264f1-9e55-43bb-bcbb-94646ee56da0 req-a580fd8c-33a9-4698-b32f-a0e9e2cf91d8 service nova] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Received event network-changed-bfc894ba-9915-48bc-b7b2-83785909f7d4 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 735.248049] env[61867]: DEBUG nova.compute.manager [req-b93264f1-9e55-43bb-bcbb-94646ee56da0 req-a580fd8c-33a9-4698-b32f-a0e9e2cf91d8 service nova] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Refreshing instance network info cache due to event network-changed-bfc894ba-9915-48bc-b7b2-83785909f7d4. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 735.248201] env[61867]: DEBUG oslo_concurrency.lockutils [req-b93264f1-9e55-43bb-bcbb-94646ee56da0 req-a580fd8c-33a9-4698-b32f-a0e9e2cf91d8 service nova] Acquiring lock "refresh_cache-5e97f142-b5c0-4318-9a35-085234d1cb64" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.248349] env[61867]: DEBUG oslo_concurrency.lockutils [req-b93264f1-9e55-43bb-bcbb-94646ee56da0 req-a580fd8c-33a9-4698-b32f-a0e9e2cf91d8 service nova] Acquired lock "refresh_cache-5e97f142-b5c0-4318-9a35-085234d1cb64" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.248508] env[61867]: DEBUG nova.network.neutron [req-b93264f1-9e55-43bb-bcbb-94646ee56da0 req-a580fd8c-33a9-4698-b32f-a0e9e2cf91d8 service nova] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Refreshing network info cache for port bfc894ba-9915-48bc-b7b2-83785909f7d4 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 735.385628] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.939s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.385628] env[61867]: ERROR nova.compute.manager [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e2d64557-ec03-4212-bfcc-6e3d8cc9f0f0, please check neutron logs for more information. [ 735.385628] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Traceback (most recent call last): [ 735.385628] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 735.385628] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] self.driver.spawn(context, instance, image_meta, [ 735.385628] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 735.385628] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 735.385628] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 735.385628] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] vm_ref = self.build_virtual_machine(instance, [ 735.386200] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 735.386200] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 735.386200] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 735.386200] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] for vif in network_info: [ 735.386200] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 735.386200] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] return self._sync_wrapper(fn, *args, **kwargs) [ 735.386200] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 735.386200] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] self.wait() [ 735.386200] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 735.386200] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] self[:] = self._gt.wait() [ 735.386200] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 735.386200] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] return self._exit_event.wait() [ 735.386200] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 735.386626] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] current.throw(*self._exc) [ 735.386626] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.386626] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] result = function(*args, **kwargs) [ 735.386626] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.386626] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] return func(*args, **kwargs) [ 735.386626] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.386626] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] raise e [ 735.386626] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.386626] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] nwinfo = self.network_api.allocate_for_instance( [ 735.386626] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.386626] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] created_port_ids = self._update_ports_for_instance( [ 735.386626] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.386626] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] with excutils.save_and_reraise_exception(): [ 735.387098] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.387098] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] self.force_reraise() [ 735.387098] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.387098] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] raise self.value [ 735.387098] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.387098] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] updated_port = self._update_port( [ 735.387098] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.387098] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] _ensure_no_port_binding_failure(port) [ 735.387098] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.387098] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] raise exception.PortBindingFailed(port_id=port['id']) [ 735.387098] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] nova.exception.PortBindingFailed: Binding failed for port e2d64557-ec03-4212-bfcc-6e3d8cc9f0f0, please check neutron logs for more information. [ 735.387098] env[61867]: ERROR nova.compute.manager [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] [ 735.387522] env[61867]: DEBUG nova.compute.utils [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Binding failed for port e2d64557-ec03-4212-bfcc-6e3d8cc9f0f0, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 735.387522] env[61867]: DEBUG oslo_concurrency.lockutils [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.733s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.389697] env[61867]: INFO nova.compute.claims [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.393235] env[61867]: DEBUG nova.compute.manager [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Build of instance 5dab4fcf-1a5e-493a-842b-b652e99d67ba was re-scheduled: Binding failed for port e2d64557-ec03-4212-bfcc-6e3d8cc9f0f0, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 735.393700] env[61867]: DEBUG nova.compute.manager [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 735.394109] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Acquiring lock "refresh_cache-5dab4fcf-1a5e-493a-842b-b652e99d67ba" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.394285] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Acquired lock "refresh_cache-5dab4fcf-1a5e-493a-842b-b652e99d67ba" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.394447] env[61867]: DEBUG nova.network.neutron [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 735.470067] env[61867]: DEBUG nova.compute.manager [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 735.479045] env[61867]: DEBUG oslo_vmware.api [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276379, 'name': PowerOnVM_Task, 'duration_secs': 0.669242} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.479045] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 735.479045] env[61867]: INFO nova.compute.manager [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Took 6.30 seconds to spawn the instance on the hypervisor. [ 735.479279] env[61867]: DEBUG nova.compute.manager [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 735.480234] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb55c00-985e-483f-85ae-17664530c9ee {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.491628] env[61867]: ERROR nova.compute.manager [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bfc894ba-9915-48bc-b7b2-83785909f7d4, please check neutron logs for more information. [ 735.491628] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 735.491628] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.491628] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 735.491628] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.491628] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 735.491628] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.491628] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 735.491628] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.491628] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 735.491628] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.491628] env[61867]: ERROR nova.compute.manager raise self.value [ 735.491628] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.491628] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 735.491628] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.491628] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 735.492198] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.492198] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 735.492198] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bfc894ba-9915-48bc-b7b2-83785909f7d4, please check neutron logs for more information. [ 735.492198] env[61867]: ERROR nova.compute.manager [ 735.492198] env[61867]: Traceback (most recent call last): [ 735.492198] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 735.492198] env[61867]: listener.cb(fileno) [ 735.492198] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.492198] env[61867]: result = function(*args, **kwargs) [ 735.492198] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.492198] env[61867]: return func(*args, **kwargs) [ 735.492198] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.492198] env[61867]: raise e [ 735.492198] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.492198] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 735.492198] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.492198] env[61867]: created_port_ids = self._update_ports_for_instance( [ 735.492198] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.492198] env[61867]: with excutils.save_and_reraise_exception(): [ 735.492198] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.492198] env[61867]: self.force_reraise() [ 735.492198] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.492198] env[61867]: raise self.value [ 735.492198] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.492198] env[61867]: updated_port = self._update_port( [ 735.492198] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.492198] env[61867]: _ensure_no_port_binding_failure(port) [ 735.492198] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.492198] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 735.493117] env[61867]: nova.exception.PortBindingFailed: Binding failed for port bfc894ba-9915-48bc-b7b2-83785909f7d4, please check neutron logs for more information. [ 735.493117] env[61867]: Removing descriptor: 21 [ 735.500272] env[61867]: DEBUG nova.virt.hardware [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 735.500510] env[61867]: DEBUG nova.virt.hardware [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 735.500664] env[61867]: DEBUG nova.virt.hardware [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.500839] env[61867]: DEBUG nova.virt.hardware [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 735.500982] env[61867]: DEBUG nova.virt.hardware [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.501158] env[61867]: DEBUG nova.virt.hardware [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 735.501400] env[61867]: DEBUG nova.virt.hardware [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 735.501580] env[61867]: DEBUG nova.virt.hardware [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 735.501752] env[61867]: DEBUG nova.virt.hardware [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 735.501912] env[61867]: DEBUG nova.virt.hardware [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 735.502097] env[61867]: DEBUG nova.virt.hardware [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.502940] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60db047e-7f75-4d06-99d3-e7ad2527920e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.511099] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7678c43c-498b-4156-9551-67003c791e12 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.527022] env[61867]: ERROR nova.compute.manager [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bfc894ba-9915-48bc-b7b2-83785909f7d4, please check neutron logs for more information. [ 735.527022] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Traceback (most recent call last): [ 735.527022] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 735.527022] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] yield resources [ 735.527022] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 735.527022] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] self.driver.spawn(context, instance, image_meta, [ 735.527022] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 735.527022] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] self._vmops.spawn(context, instance, image_meta, injected_files, [ 735.527022] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 735.527022] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] vm_ref = self.build_virtual_machine(instance, [ 735.527022] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 735.527480] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] vif_infos = vmwarevif.get_vif_info(self._session, [ 735.527480] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 735.527480] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] for vif in network_info: [ 735.527480] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 735.527480] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] return self._sync_wrapper(fn, *args, **kwargs) [ 735.527480] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 735.527480] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] self.wait() [ 735.527480] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 735.527480] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] self[:] = self._gt.wait() [ 735.527480] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 735.527480] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] return self._exit_event.wait() [ 735.527480] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 735.527480] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] current.throw(*self._exc) [ 735.527851] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.527851] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] result = function(*args, **kwargs) [ 735.527851] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.527851] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] return func(*args, **kwargs) [ 735.527851] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.527851] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] raise e [ 735.527851] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.527851] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] nwinfo = self.network_api.allocate_for_instance( [ 735.527851] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.527851] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] created_port_ids = self._update_ports_for_instance( [ 735.527851] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.527851] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] with excutils.save_and_reraise_exception(): [ 735.527851] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.528254] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] self.force_reraise() [ 735.528254] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.528254] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] raise self.value [ 735.528254] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.528254] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] updated_port = self._update_port( [ 735.528254] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.528254] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] _ensure_no_port_binding_failure(port) [ 735.528254] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.528254] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] raise exception.PortBindingFailed(port_id=port['id']) [ 735.528254] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] nova.exception.PortBindingFailed: Binding failed for port bfc894ba-9915-48bc-b7b2-83785909f7d4, please check neutron logs for more information. [ 735.528254] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] [ 735.528254] env[61867]: INFO nova.compute.manager [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Terminating instance [ 735.528663] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Acquiring lock "refresh_cache-5e97f142-b5c0-4318-9a35-085234d1cb64" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.735970] env[61867]: DEBUG nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 735.769895] env[61867]: DEBUG nova.network.neutron [req-b93264f1-9e55-43bb-bcbb-94646ee56da0 req-a580fd8c-33a9-4698-b32f-a0e9e2cf91d8 service nova] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.818415] env[61867]: DEBUG nova.network.neutron [req-b93264f1-9e55-43bb-bcbb-94646ee56da0 req-a580fd8c-33a9-4698-b32f-a0e9e2cf91d8 service nova] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.913901] env[61867]: DEBUG nova.network.neutron [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.958613] env[61867]: DEBUG nova.network.neutron [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.002456] env[61867]: INFO nova.compute.manager [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Took 29.52 seconds to build instance. [ 736.257994] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.321589] env[61867]: DEBUG oslo_concurrency.lockutils [req-b93264f1-9e55-43bb-bcbb-94646ee56da0 req-a580fd8c-33a9-4698-b32f-a0e9e2cf91d8 service nova] Releasing lock "refresh_cache-5e97f142-b5c0-4318-9a35-085234d1cb64" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.321589] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Acquired lock "refresh_cache-5e97f142-b5c0-4318-9a35-085234d1cb64" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.321589] env[61867]: DEBUG nova.network.neutron [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 736.461903] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Releasing lock "refresh_cache-5dab4fcf-1a5e-493a-842b-b652e99d67ba" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.462450] env[61867]: DEBUG nova.compute.manager [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 736.462730] env[61867]: DEBUG nova.compute.manager [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 736.462922] env[61867]: DEBUG nova.network.neutron [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 736.483396] env[61867]: DEBUG nova.network.neutron [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.504872] env[61867]: DEBUG oslo_concurrency.lockutils [None req-24526beb-796b-4faa-b4fb-5c432bfd4ab9 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "8656c3b8-b7b1-442a-9719-1588ee83d19b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.405s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.674706] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c26b608-c9fd-4ceb-8ddb-fa6cee0c9123 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.682273] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d41a1aa-71ad-4cc7-a577-d64be33d14c1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.711206] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c871726-9a5f-4274-82de-a9f15929e8aa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.717523] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07de0c82-ee47-4843-82bb-f00831b78868 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.729639] env[61867]: DEBUG nova.compute.provider_tree [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.891558] env[61867]: INFO nova.compute.manager [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Rebuilding instance [ 736.930897] env[61867]: DEBUG nova.compute.manager [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 736.931876] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7bf7fd-845d-4d4d-b084-82d482dbe470 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.979302] env[61867]: DEBUG nova.network.neutron [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.985735] env[61867]: DEBUG nova.network.neutron [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.007338] env[61867]: DEBUG nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 737.041768] env[61867]: DEBUG nova.network.neutron [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.232472] env[61867]: DEBUG nova.scheduler.client.report [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.273958] env[61867]: DEBUG nova.compute.manager [req-f423d0f7-1996-44c8-8330-7b10e7af951c req-95bc39f4-e70d-47f6-8788-1bb91fd45dfb service nova] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Received event network-vif-deleted-bfc894ba-9915-48bc-b7b2-83785909f7d4 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.443155] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 737.443501] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-236dce78-a933-4f3b-89a4-47fd33750f5b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.451864] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 737.451864] env[61867]: value = "task-1276380" [ 737.451864] env[61867]: _type = "Task" [ 737.451864] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.461423] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276380, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.490034] env[61867]: INFO nova.compute.manager [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] [instance: 5dab4fcf-1a5e-493a-842b-b652e99d67ba] Took 1.03 seconds to deallocate network for instance. [ 737.535475] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.544452] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Releasing lock "refresh_cache-5e97f142-b5c0-4318-9a35-085234d1cb64" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.544894] env[61867]: DEBUG nova.compute.manager [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 737.545171] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 737.545624] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b5766190-9b47-426f-a288-991771c6bc43 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.554692] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50779a18-6208-4562-b9ce-24539c3e0317 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.580351] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5e97f142-b5c0-4318-9a35-085234d1cb64 could not be found. [ 737.580351] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 737.580351] env[61867]: INFO nova.compute.manager [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Took 0.03 seconds to destroy the instance on the hypervisor. [ 737.580351] env[61867]: DEBUG oslo.service.loopingcall [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.580351] env[61867]: DEBUG nova.compute.manager [-] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 737.580351] env[61867]: DEBUG nova.network.neutron [-] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 737.597152] env[61867]: DEBUG nova.network.neutron [-] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.737256] env[61867]: DEBUG oslo_concurrency.lockutils [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.737804] env[61867]: DEBUG nova.compute.manager [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 737.740732] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.379s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.961414] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276380, 'name': PowerOffVM_Task, 'duration_secs': 0.197624} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.961669] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 737.961880] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 737.962652] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fadc5d37-a3ce-4601-ad82-2e07c0b4e1d5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.969259] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 737.969367] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a5868c64-8985-4ca5-9af7-7af648040b20 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.992256] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 737.992471] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 737.992651] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Deleting the datastore file [datastore2] 8656c3b8-b7b1-442a-9719-1588ee83d19b {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 737.992892] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-712c83c6-b646-44df-9778-8eafd44cf44d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.000901] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 738.000901] env[61867]: value = "task-1276382" [ 738.000901] env[61867]: _type = "Task" [ 738.000901] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.008685] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276382, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.102541] env[61867]: DEBUG nova.network.neutron [-] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.249378] env[61867]: DEBUG nova.compute.utils [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 738.250905] env[61867]: DEBUG nova.compute.manager [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 738.251088] env[61867]: DEBUG nova.network.neutron [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 738.300561] env[61867]: DEBUG nova.policy [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a09cf647871344129f9c88e39e89a11e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '821b762613384c5b841b44a64d8a6f28', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 738.515775] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276382, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.519975] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c19a83c8-462f-464f-82e4-e840ca28e151 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.522509] env[61867]: INFO nova.scheduler.client.report [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Deleted allocations for instance 5dab4fcf-1a5e-493a-842b-b652e99d67ba [ 738.535543] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fc2aefd-a03f-4766-9fbe-2a094792d81a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.571239] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8040b8-dd28-4d49-9b42-edcb1df287b5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.579061] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b14295f-97bd-4822-b860-23607fba36e6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.592950] env[61867]: DEBUG nova.compute.provider_tree [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.605522] env[61867]: INFO nova.compute.manager [-] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Took 1.03 seconds to deallocate network for instance. [ 738.611483] env[61867]: DEBUG nova.compute.claims [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 738.611483] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.654216] env[61867]: DEBUG nova.network.neutron [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Successfully created port: 4c609ff8-9b47-4769-ba7e-0618fe97fd29 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 738.754634] env[61867]: DEBUG nova.compute.manager [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 739.015183] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276382, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.528979} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.015514] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 739.015722] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 739.015939] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 739.035166] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2d018205-2f8a-430e-9bc2-30126f97aeb3 tempest-ServerDiagnosticsTest-839117532 tempest-ServerDiagnosticsTest-839117532-project-member] Lock "5dab4fcf-1a5e-493a-842b-b652e99d67ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.933s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.096240] env[61867]: DEBUG nova.scheduler.client.report [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.302688] env[61867]: DEBUG nova.compute.manager [req-dd5d5378-264e-4f51-83a2-3e5e2278b10f req-089878a8-0286-4d97-ace7-a4f787a1d0ff service nova] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Received event network-changed-4c609ff8-9b47-4769-ba7e-0618fe97fd29 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 739.302915] env[61867]: DEBUG nova.compute.manager [req-dd5d5378-264e-4f51-83a2-3e5e2278b10f req-089878a8-0286-4d97-ace7-a4f787a1d0ff service nova] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Refreshing instance network info cache due to event network-changed-4c609ff8-9b47-4769-ba7e-0618fe97fd29. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 739.303259] env[61867]: DEBUG oslo_concurrency.lockutils [req-dd5d5378-264e-4f51-83a2-3e5e2278b10f req-089878a8-0286-4d97-ace7-a4f787a1d0ff service nova] Acquiring lock "refresh_cache-796ab4d3-ea9a-495d-9807-7780b1a7b0d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.303441] env[61867]: DEBUG oslo_concurrency.lockutils [req-dd5d5378-264e-4f51-83a2-3e5e2278b10f req-089878a8-0286-4d97-ace7-a4f787a1d0ff service nova] Acquired lock "refresh_cache-796ab4d3-ea9a-495d-9807-7780b1a7b0d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.303564] env[61867]: DEBUG nova.network.neutron [req-dd5d5378-264e-4f51-83a2-3e5e2278b10f req-089878a8-0286-4d97-ace7-a4f787a1d0ff service nova] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Refreshing network info cache for port 4c609ff8-9b47-4769-ba7e-0618fe97fd29 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 739.458912] env[61867]: ERROR nova.compute.manager [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4c609ff8-9b47-4769-ba7e-0618fe97fd29, please check neutron logs for more information. [ 739.458912] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 739.458912] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.458912] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 739.458912] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.458912] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 739.458912] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.458912] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 739.458912] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.458912] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 739.458912] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.458912] env[61867]: ERROR nova.compute.manager raise self.value [ 739.458912] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.458912] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 739.458912] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.458912] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 739.459662] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.459662] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 739.459662] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4c609ff8-9b47-4769-ba7e-0618fe97fd29, please check neutron logs for more information. [ 739.459662] env[61867]: ERROR nova.compute.manager [ 739.459662] env[61867]: Traceback (most recent call last): [ 739.459662] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 739.459662] env[61867]: listener.cb(fileno) [ 739.459662] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.459662] env[61867]: result = function(*args, **kwargs) [ 739.459662] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 739.459662] env[61867]: return func(*args, **kwargs) [ 739.459662] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.459662] env[61867]: raise e [ 739.459662] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.459662] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 739.459662] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.459662] env[61867]: created_port_ids = self._update_ports_for_instance( [ 739.459662] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.459662] env[61867]: with excutils.save_and_reraise_exception(): [ 739.459662] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.459662] env[61867]: self.force_reraise() [ 739.459662] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.459662] env[61867]: raise self.value [ 739.459662] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.459662] env[61867]: updated_port = self._update_port( [ 739.459662] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.459662] env[61867]: _ensure_no_port_binding_failure(port) [ 739.459662] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.459662] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 739.460795] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 4c609ff8-9b47-4769-ba7e-0618fe97fd29, please check neutron logs for more information. [ 739.460795] env[61867]: Removing descriptor: 21 [ 739.540088] env[61867]: DEBUG nova.compute.manager [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 739.602245] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.861s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.603017] env[61867]: ERROR nova.compute.manager [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d1b94eee-4eda-46c9-b144-f5356b51897e, please check neutron logs for more information. [ 739.603017] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Traceback (most recent call last): [ 739.603017] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 739.603017] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] self.driver.spawn(context, instance, image_meta, [ 739.603017] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 739.603017] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.603017] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.603017] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] vm_ref = self.build_virtual_machine(instance, [ 739.603017] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.603017] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.603017] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.603391] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] for vif in network_info: [ 739.603391] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.603391] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] return self._sync_wrapper(fn, *args, **kwargs) [ 739.603391] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.603391] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] self.wait() [ 739.603391] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.603391] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] self[:] = self._gt.wait() [ 739.603391] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.603391] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] return self._exit_event.wait() [ 739.603391] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 739.603391] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] current.throw(*self._exc) [ 739.603391] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.603391] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] result = function(*args, **kwargs) [ 739.604120] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 739.604120] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] return func(*args, **kwargs) [ 739.604120] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.604120] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] raise e [ 739.604120] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.604120] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] nwinfo = self.network_api.allocate_for_instance( [ 739.604120] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.604120] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] created_port_ids = self._update_ports_for_instance( [ 739.604120] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.604120] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] with excutils.save_and_reraise_exception(): [ 739.604120] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.604120] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] self.force_reraise() [ 739.604120] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.604557] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] raise self.value [ 739.604557] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.604557] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] updated_port = self._update_port( [ 739.604557] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.604557] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] _ensure_no_port_binding_failure(port) [ 739.604557] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.604557] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] raise exception.PortBindingFailed(port_id=port['id']) [ 739.604557] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] nova.exception.PortBindingFailed: Binding failed for port d1b94eee-4eda-46c9-b144-f5356b51897e, please check neutron logs for more information. [ 739.604557] env[61867]: ERROR nova.compute.manager [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] [ 739.604557] env[61867]: DEBUG nova.compute.utils [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Binding failed for port d1b94eee-4eda-46c9-b144-f5356b51897e, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 739.604895] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.593s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.606273] env[61867]: INFO nova.compute.claims [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 739.608957] env[61867]: DEBUG nova.compute.manager [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Build of instance ab18091f-bab5-4926-b7a0-1f0e005bebe0 was re-scheduled: Binding failed for port d1b94eee-4eda-46c9-b144-f5356b51897e, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 739.609555] env[61867]: DEBUG nova.compute.manager [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 739.609780] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "refresh_cache-ab18091f-bab5-4926-b7a0-1f0e005bebe0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.609927] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquired lock "refresh_cache-ab18091f-bab5-4926-b7a0-1f0e005bebe0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.610101] env[61867]: DEBUG nova.network.neutron [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 739.771497] env[61867]: DEBUG nova.compute.manager [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 739.795752] env[61867]: DEBUG nova.virt.hardware [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 739.795752] env[61867]: DEBUG nova.virt.hardware [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 739.795946] env[61867]: DEBUG nova.virt.hardware [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 739.796043] env[61867]: DEBUG nova.virt.hardware [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 739.796194] env[61867]: DEBUG nova.virt.hardware [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 739.796336] env[61867]: DEBUG nova.virt.hardware [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 739.796540] env[61867]: DEBUG nova.virt.hardware [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 739.796696] env[61867]: DEBUG nova.virt.hardware [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 739.796857] env[61867]: DEBUG nova.virt.hardware [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 739.797027] env[61867]: DEBUG nova.virt.hardware [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 739.797211] env[61867]: DEBUG nova.virt.hardware [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 739.798070] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa0bef2-9050-4cc6-b931-759dbb5e47cb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.806116] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef7fef69-7103-4e73-bae5-2c9122286ce9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.823192] env[61867]: ERROR nova.compute.manager [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4c609ff8-9b47-4769-ba7e-0618fe97fd29, please check neutron logs for more information. [ 739.823192] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Traceback (most recent call last): [ 739.823192] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 739.823192] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] yield resources [ 739.823192] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 739.823192] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] self.driver.spawn(context, instance, image_meta, [ 739.823192] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 739.823192] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.823192] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.823192] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] vm_ref = self.build_virtual_machine(instance, [ 739.823192] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.823630] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.823630] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.823630] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] for vif in network_info: [ 739.823630] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.823630] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] return self._sync_wrapper(fn, *args, **kwargs) [ 739.823630] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.823630] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] self.wait() [ 739.823630] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.823630] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] self[:] = self._gt.wait() [ 739.823630] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.823630] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] return self._exit_event.wait() [ 739.823630] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 739.823630] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] current.throw(*self._exc) [ 739.824124] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.824124] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] result = function(*args, **kwargs) [ 739.824124] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 739.824124] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] return func(*args, **kwargs) [ 739.824124] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.824124] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] raise e [ 739.824124] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.824124] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] nwinfo = self.network_api.allocate_for_instance( [ 739.824124] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.824124] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] created_port_ids = self._update_ports_for_instance( [ 739.824124] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.824124] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] with excutils.save_and_reraise_exception(): [ 739.824124] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.824590] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] self.force_reraise() [ 739.824590] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.824590] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] raise self.value [ 739.824590] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.824590] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] updated_port = self._update_port( [ 739.824590] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.824590] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] _ensure_no_port_binding_failure(port) [ 739.824590] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.824590] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] raise exception.PortBindingFailed(port_id=port['id']) [ 739.824590] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] nova.exception.PortBindingFailed: Binding failed for port 4c609ff8-9b47-4769-ba7e-0618fe97fd29, please check neutron logs for more information. [ 739.824590] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] [ 739.824590] env[61867]: INFO nova.compute.manager [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Terminating instance [ 739.825596] env[61867]: DEBUG nova.network.neutron [req-dd5d5378-264e-4f51-83a2-3e5e2278b10f req-089878a8-0286-4d97-ace7-a4f787a1d0ff service nova] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.827696] env[61867]: DEBUG oslo_concurrency.lockutils [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Acquiring lock "refresh_cache-796ab4d3-ea9a-495d-9807-7780b1a7b0d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.876485] env[61867]: DEBUG nova.network.neutron [req-dd5d5378-264e-4f51-83a2-3e5e2278b10f req-089878a8-0286-4d97-ace7-a4f787a1d0ff service nova] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.051519] env[61867]: DEBUG nova.virt.hardware [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 740.051777] env[61867]: DEBUG nova.virt.hardware [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 740.051937] env[61867]: DEBUG nova.virt.hardware [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 740.052117] env[61867]: DEBUG nova.virt.hardware [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 740.052268] env[61867]: DEBUG nova.virt.hardware [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 740.055453] env[61867]: DEBUG nova.virt.hardware [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 740.055453] env[61867]: DEBUG nova.virt.hardware [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 740.055453] env[61867]: DEBUG nova.virt.hardware [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 740.055453] env[61867]: DEBUG nova.virt.hardware [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 740.055453] env[61867]: DEBUG nova.virt.hardware [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 740.055741] env[61867]: DEBUG nova.virt.hardware [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 740.055741] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eccc23e-5d7c-41f0-aacb-b64e44112ea6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.062066] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2aaad7b-c27f-4ff8-bcae-8af08779d732 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.066655] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.076548] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Instance VIF info [] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 740.084101] env[61867]: DEBUG oslo.service.loopingcall [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.084101] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 740.084101] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d32418f2-559f-4d44-9f9c-fe4912b63b35 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.099858] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 740.099858] env[61867]: value = "task-1276383" [ 740.099858] env[61867]: _type = "Task" [ 740.099858] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.109100] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276383, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.137461] env[61867]: DEBUG nova.network.neutron [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.280466] env[61867]: DEBUG nova.network.neutron [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.379918] env[61867]: DEBUG oslo_concurrency.lockutils [req-dd5d5378-264e-4f51-83a2-3e5e2278b10f req-089878a8-0286-4d97-ace7-a4f787a1d0ff service nova] Releasing lock "refresh_cache-796ab4d3-ea9a-495d-9807-7780b1a7b0d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.379918] env[61867]: DEBUG oslo_concurrency.lockutils [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Acquired lock "refresh_cache-796ab4d3-ea9a-495d-9807-7780b1a7b0d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.380195] env[61867]: DEBUG nova.network.neutron [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 740.609200] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276383, 'name': CreateVM_Task, 'duration_secs': 0.283706} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.609506] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 740.609831] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.609988] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.610361] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 740.610574] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71b483a1-681c-4f6e-9a63-4d32cd2a15a1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.614848] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 740.614848] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5278faaa-03af-5398-2355-add30967337b" [ 740.614848] env[61867]: _type = "Task" [ 740.614848] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.625569] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5278faaa-03af-5398-2355-add30967337b, 'name': SearchDatastore_Task, 'duration_secs': 0.008204} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.625670] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.625847] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 740.626081] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.626224] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.626399] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 740.626634] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f6088b11-85f9-4482-895e-a1c79c6244b9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.633263] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 740.633448] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 740.634118] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67f74ae3-0249-45cf-bedd-faef09159769 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.641026] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 740.641026] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5241edd5-55f6-b94f-81c8-a058ca3db965" [ 740.641026] env[61867]: _type = "Task" [ 740.641026] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.648248] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5241edd5-55f6-b94f-81c8-a058ca3db965, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.783467] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Releasing lock "refresh_cache-ab18091f-bab5-4926-b7a0-1f0e005bebe0" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.783711] env[61867]: DEBUG nova.compute.manager [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 740.783892] env[61867]: DEBUG nova.compute.manager [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.784070] env[61867]: DEBUG nova.network.neutron [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.802507] env[61867]: DEBUG nova.network.neutron [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.891040] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b8636d0-3659-45f8-9a88-e43e7e6f5dde {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.899591] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0794be5d-0caa-4f62-b25f-33c3ef6eb3a3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.901390] env[61867]: DEBUG nova.network.neutron [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.931100] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa95870-f943-4cdc-9cf2-d576fb94ee8f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.938772] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6909983a-f610-4afe-8cd8-6e35be9cd469 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.953368] env[61867]: DEBUG nova.compute.provider_tree [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.986851] env[61867]: DEBUG nova.network.neutron [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.151260] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5241edd5-55f6-b94f-81c8-a058ca3db965, 'name': SearchDatastore_Task, 'duration_secs': 0.007144} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.152071] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7df629fb-eede-486f-89d4-118948901605 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.157246] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 741.157246] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b9be94-667c-67d4-3269-c88e29975f50" [ 741.157246] env[61867]: _type = "Task" [ 741.157246] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.165773] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b9be94-667c-67d4-3269-c88e29975f50, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.305328] env[61867]: DEBUG nova.network.neutron [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.328354] env[61867]: DEBUG nova.compute.manager [req-84374ee6-f552-4610-9310-206d7576e9cc req-8266d905-de78-4e9d-98b6-d896facbd648 service nova] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Received event network-vif-deleted-4c609ff8-9b47-4769-ba7e-0618fe97fd29 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 741.459086] env[61867]: DEBUG nova.scheduler.client.report [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.489949] env[61867]: DEBUG oslo_concurrency.lockutils [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Releasing lock "refresh_cache-796ab4d3-ea9a-495d-9807-7780b1a7b0d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.490486] env[61867]: DEBUG nova.compute.manager [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 741.490792] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 741.491360] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c16470e1-cdd9-4b25-bac4-e25c084cf360 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.500911] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7cacf7-6b8c-470a-b7fc-5fc105510b93 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.523958] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 796ab4d3-ea9a-495d-9807-7780b1a7b0d6 could not be found. [ 741.524209] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 741.524393] env[61867]: INFO nova.compute.manager [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 741.524689] env[61867]: DEBUG oslo.service.loopingcall [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 741.524915] env[61867]: DEBUG nova.compute.manager [-] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 741.525037] env[61867]: DEBUG nova.network.neutron [-] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 741.543920] env[61867]: DEBUG nova.network.neutron [-] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.667352] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b9be94-667c-67d4-3269-c88e29975f50, 'name': SearchDatastore_Task, 'duration_secs': 0.009257} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.667625] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.667887] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 8656c3b8-b7b1-442a-9719-1588ee83d19b/8656c3b8-b7b1-442a-9719-1588ee83d19b.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 741.668149] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-64d8ee4f-48de-4b28-b34c-3140432917cb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.674123] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 741.674123] env[61867]: value = "task-1276384" [ 741.674123] env[61867]: _type = "Task" [ 741.674123] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.681015] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276384, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.811024] env[61867]: INFO nova.compute.manager [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: ab18091f-bab5-4926-b7a0-1f0e005bebe0] Took 1.02 seconds to deallocate network for instance. [ 741.964433] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.359s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.966947] env[61867]: DEBUG nova.compute.manager [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 741.975655] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.044s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.978806] env[61867]: INFO nova.compute.claims [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 742.045185] env[61867]: DEBUG nova.network.neutron [-] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.185698] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276384, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448039} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.185958] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 8656c3b8-b7b1-442a-9719-1588ee83d19b/8656c3b8-b7b1-442a-9719-1588ee83d19b.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 742.186178] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 742.186794] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-71a006fa-5991-40e3-9775-49579204f03d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.192744] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 742.192744] env[61867]: value = "task-1276385" [ 742.192744] env[61867]: _type = "Task" [ 742.192744] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.200529] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276385, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.487194] env[61867]: DEBUG nova.compute.utils [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 742.487194] env[61867]: DEBUG nova.compute.manager [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 742.487194] env[61867]: DEBUG nova.network.neutron [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 742.548536] env[61867]: INFO nova.compute.manager [-] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Took 1.02 seconds to deallocate network for instance. [ 742.552845] env[61867]: DEBUG nova.compute.claims [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 742.552845] env[61867]: DEBUG oslo_concurrency.lockutils [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.564077] env[61867]: DEBUG nova.policy [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a09cf647871344129f9c88e39e89a11e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '821b762613384c5b841b44a64d8a6f28', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 742.702844] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276385, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055035} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.703271] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 742.704018] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc9c9e5-bfa7-470a-ad86-ef8530f62312 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.722701] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] 8656c3b8-b7b1-442a-9719-1588ee83d19b/8656c3b8-b7b1-442a-9719-1588ee83d19b.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 742.723110] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7fde94d-8fa1-4141-8867-396848c523a2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.743274] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 742.743274] env[61867]: value = "task-1276386" [ 742.743274] env[61867]: _type = "Task" [ 742.743274] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.751034] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276386, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.809364] env[61867]: DEBUG nova.network.neutron [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Successfully created port: 35552c32-4d31-4fa5-a23d-e2dc0035d263 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 742.851212] env[61867]: INFO nova.scheduler.client.report [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Deleted allocations for instance ab18091f-bab5-4926-b7a0-1f0e005bebe0 [ 742.995889] env[61867]: DEBUG nova.compute.manager [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 743.254097] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276386, 'name': ReconfigVM_Task, 'duration_secs': 0.272974} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.254305] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Reconfigured VM instance instance-0000002f to attach disk [datastore2] 8656c3b8-b7b1-442a-9719-1588ee83d19b/8656c3b8-b7b1-442a-9719-1588ee83d19b.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 743.255798] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-61cfc929-2155-455b-b7a3-a696999437a8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.263441] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 743.263441] env[61867]: value = "task-1276387" [ 743.263441] env[61867]: _type = "Task" [ 743.263441] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.274236] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276387, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.362925] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6595565b-3c62-422a-a308-f06702105423 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "ab18091f-bab5-4926-b7a0-1f0e005bebe0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.843s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.371326] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-194dc14f-4342-46a7-8fde-e21bae8fdb10 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.382671] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eda72e1-7dee-4b5f-bed2-3057ae45a2df {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.418561] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbbcf69e-b808-4dd0-8f39-beca0bfd95c5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.426162] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5145ec73-1afc-4f01-b99a-279986c558c6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.440811] env[61867]: DEBUG nova.compute.provider_tree [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.708627] env[61867]: DEBUG nova.compute.manager [req-0343ddad-02a3-489b-9862-84c86a6662c8 req-6767ebba-0e47-46c4-900b-ca695d194f3b service nova] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Received event network-changed-35552c32-4d31-4fa5-a23d-e2dc0035d263 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 743.708840] env[61867]: DEBUG nova.compute.manager [req-0343ddad-02a3-489b-9862-84c86a6662c8 req-6767ebba-0e47-46c4-900b-ca695d194f3b service nova] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Refreshing instance network info cache due to event network-changed-35552c32-4d31-4fa5-a23d-e2dc0035d263. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 743.708983] env[61867]: DEBUG oslo_concurrency.lockutils [req-0343ddad-02a3-489b-9862-84c86a6662c8 req-6767ebba-0e47-46c4-900b-ca695d194f3b service nova] Acquiring lock "refresh_cache-8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.709204] env[61867]: DEBUG oslo_concurrency.lockutils [req-0343ddad-02a3-489b-9862-84c86a6662c8 req-6767ebba-0e47-46c4-900b-ca695d194f3b service nova] Acquired lock "refresh_cache-8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.709403] env[61867]: DEBUG nova.network.neutron [req-0343ddad-02a3-489b-9862-84c86a6662c8 req-6767ebba-0e47-46c4-900b-ca695d194f3b service nova] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Refreshing network info cache for port 35552c32-4d31-4fa5-a23d-e2dc0035d263 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 743.779909] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276387, 'name': Rename_Task, 'duration_secs': 0.136218} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.780710] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 743.781698] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-232b43b9-b984-442b-9cfd-92162766c2fa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.788368] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 743.788368] env[61867]: value = "task-1276388" [ 743.788368] env[61867]: _type = "Task" [ 743.788368] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.796152] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276388, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.808353] env[61867]: ERROR nova.compute.manager [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 35552c32-4d31-4fa5-a23d-e2dc0035d263, please check neutron logs for more information. [ 743.808353] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 743.808353] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.808353] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 743.808353] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.808353] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 743.808353] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.808353] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 743.808353] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.808353] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 743.808353] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.808353] env[61867]: ERROR nova.compute.manager raise self.value [ 743.808353] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.808353] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 743.808353] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.808353] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 743.808861] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.808861] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 743.808861] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 35552c32-4d31-4fa5-a23d-e2dc0035d263, please check neutron logs for more information. [ 743.808861] env[61867]: ERROR nova.compute.manager [ 743.808861] env[61867]: Traceback (most recent call last): [ 743.808861] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 743.808861] env[61867]: listener.cb(fileno) [ 743.808861] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.808861] env[61867]: result = function(*args, **kwargs) [ 743.808861] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 743.808861] env[61867]: return func(*args, **kwargs) [ 743.808861] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.808861] env[61867]: raise e [ 743.808861] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.808861] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 743.808861] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.808861] env[61867]: created_port_ids = self._update_ports_for_instance( [ 743.808861] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.808861] env[61867]: with excutils.save_and_reraise_exception(): [ 743.808861] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.808861] env[61867]: self.force_reraise() [ 743.808861] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.808861] env[61867]: raise self.value [ 743.808861] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.808861] env[61867]: updated_port = self._update_port( [ 743.808861] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.808861] env[61867]: _ensure_no_port_binding_failure(port) [ 743.808861] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.808861] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 743.809997] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 35552c32-4d31-4fa5-a23d-e2dc0035d263, please check neutron logs for more information. [ 743.809997] env[61867]: Removing descriptor: 21 [ 743.868657] env[61867]: DEBUG nova.compute.manager [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 743.943755] env[61867]: DEBUG nova.scheduler.client.report [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 744.009937] env[61867]: DEBUG nova.compute.manager [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 744.047895] env[61867]: DEBUG nova.virt.hardware [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 744.048164] env[61867]: DEBUG nova.virt.hardware [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 744.048318] env[61867]: DEBUG nova.virt.hardware [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 744.048499] env[61867]: DEBUG nova.virt.hardware [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 744.048680] env[61867]: DEBUG nova.virt.hardware [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 744.048794] env[61867]: DEBUG nova.virt.hardware [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 744.049033] env[61867]: DEBUG nova.virt.hardware [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 744.049221] env[61867]: DEBUG nova.virt.hardware [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 744.049492] env[61867]: DEBUG nova.virt.hardware [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 744.049662] env[61867]: DEBUG nova.virt.hardware [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 744.049829] env[61867]: DEBUG nova.virt.hardware [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 744.050695] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08901520-6b63-4b00-99c4-3c58c3a97350 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.058539] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7e874d-610b-440e-8d78-546dffca3515 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.072077] env[61867]: ERROR nova.compute.manager [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 35552c32-4d31-4fa5-a23d-e2dc0035d263, please check neutron logs for more information. [ 744.072077] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Traceback (most recent call last): [ 744.072077] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 744.072077] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] yield resources [ 744.072077] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 744.072077] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] self.driver.spawn(context, instance, image_meta, [ 744.072077] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 744.072077] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] self._vmops.spawn(context, instance, image_meta, injected_files, [ 744.072077] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 744.072077] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] vm_ref = self.build_virtual_machine(instance, [ 744.072077] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 744.072477] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] vif_infos = vmwarevif.get_vif_info(self._session, [ 744.072477] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 744.072477] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] for vif in network_info: [ 744.072477] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 744.072477] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] return self._sync_wrapper(fn, *args, **kwargs) [ 744.072477] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 744.072477] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] self.wait() [ 744.072477] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 744.072477] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] self[:] = self._gt.wait() [ 744.072477] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 744.072477] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] return self._exit_event.wait() [ 744.072477] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 744.072477] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] current.throw(*self._exc) [ 744.072880] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.072880] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] result = function(*args, **kwargs) [ 744.072880] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 744.072880] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] return func(*args, **kwargs) [ 744.072880] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 744.072880] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] raise e [ 744.072880] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.072880] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] nwinfo = self.network_api.allocate_for_instance( [ 744.072880] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.072880] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] created_port_ids = self._update_ports_for_instance( [ 744.072880] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.072880] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] with excutils.save_and_reraise_exception(): [ 744.072880] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.073309] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] self.force_reraise() [ 744.073309] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.073309] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] raise self.value [ 744.073309] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.073309] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] updated_port = self._update_port( [ 744.073309] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.073309] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] _ensure_no_port_binding_failure(port) [ 744.073309] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.073309] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] raise exception.PortBindingFailed(port_id=port['id']) [ 744.073309] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] nova.exception.PortBindingFailed: Binding failed for port 35552c32-4d31-4fa5-a23d-e2dc0035d263, please check neutron logs for more information. [ 744.073309] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] [ 744.073309] env[61867]: INFO nova.compute.manager [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Terminating instance [ 744.074675] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Acquiring lock "refresh_cache-8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.234414] env[61867]: DEBUG nova.network.neutron [req-0343ddad-02a3-489b-9862-84c86a6662c8 req-6767ebba-0e47-46c4-900b-ca695d194f3b service nova] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.287643] env[61867]: DEBUG nova.network.neutron [req-0343ddad-02a3-489b-9862-84c86a6662c8 req-6767ebba-0e47-46c4-900b-ca695d194f3b service nova] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.300570] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276388, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.339347] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "6e41989e-b8fa-4009-af1e-1ce859b329a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.339519] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "6e41989e-b8fa-4009-af1e-1ce859b329a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.390700] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.450228] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.476s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.450786] env[61867]: DEBUG nova.compute.manager [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 744.453443] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.943s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.795390] env[61867]: DEBUG oslo_concurrency.lockutils [req-0343ddad-02a3-489b-9862-84c86a6662c8 req-6767ebba-0e47-46c4-900b-ca695d194f3b service nova] Releasing lock "refresh_cache-8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.796498] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Acquired lock "refresh_cache-8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.797483] env[61867]: DEBUG nova.network.neutron [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 744.803912] env[61867]: DEBUG oslo_vmware.api [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276388, 'name': PowerOnVM_Task, 'duration_secs': 0.537434} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.804042] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 744.804673] env[61867]: DEBUG nova.compute.manager [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 744.805353] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f90ae50-de20-4e1b-b82f-279d6401db90 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.831515] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "8a83f4f2-58eb-473e-9b1e-32ce633554f9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.831870] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "8a83f4f2-58eb-473e-9b1e-32ce633554f9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.958044] env[61867]: DEBUG nova.compute.utils [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 744.959503] env[61867]: DEBUG nova.compute.manager [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 744.959728] env[61867]: DEBUG nova.network.neutron [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 745.037058] env[61867]: DEBUG nova.policy [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ea634246436422a9407f82692b69ede', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '86ced20eaf4740e298dc6f8ca5550c09', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 745.328210] env[61867]: DEBUG nova.network.neutron [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.330686] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.433183] env[61867]: DEBUG nova.network.neutron [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.466015] env[61867]: DEBUG nova.compute.manager [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 745.497172] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance b1ec32c5-0642-4dca-ad43-05a5172d04d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 745.497330] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 8656c3b8-b7b1-442a-9719-1588ee83d19b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 745.497446] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 5e97f142-b5c0-4318-9a35-085234d1cb64 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 745.497559] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 796ab4d3-ea9a-495d-9807-7780b1a7b0d6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 745.497682] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 745.497787] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 745.568399] env[61867]: DEBUG nova.network.neutron [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Successfully created port: 3b1112dd-2d7d-453e-ad9e-677480ae22c4 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.734063] env[61867]: DEBUG nova.compute.manager [req-dada38ef-2417-47ce-b323-2360b2a74d25 req-4a12ddb0-2d94-4183-94ee-c33af0412917 service nova] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Received event network-vif-deleted-35552c32-4d31-4fa5-a23d-e2dc0035d263 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 745.939019] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Releasing lock "refresh_cache-8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.939019] env[61867]: DEBUG nova.compute.manager [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 745.939019] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 745.939019] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d62eefc3-2f10-42e4-abf0-6006899da4bf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.947875] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a27e9fa-ecf3-4331-8dd9-bdce0fcbea42 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.981633] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce could not be found. [ 745.981941] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 745.982371] env[61867]: INFO nova.compute.manager [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Took 0.05 seconds to destroy the instance on the hypervisor. [ 745.982687] env[61867]: DEBUG oslo.service.loopingcall [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 745.983328] env[61867]: DEBUG nova.compute.manager [-] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 745.983932] env[61867]: DEBUG nova.network.neutron [-] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 746.001184] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 746.003273] env[61867]: DEBUG nova.network.neutron [-] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.117035] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "8656c3b8-b7b1-442a-9719-1588ee83d19b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.117360] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "8656c3b8-b7b1-442a-9719-1588ee83d19b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.117579] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "8656c3b8-b7b1-442a-9719-1588ee83d19b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.117762] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "8656c3b8-b7b1-442a-9719-1588ee83d19b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.117926] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "8656c3b8-b7b1-442a-9719-1588ee83d19b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.121195] env[61867]: INFO nova.compute.manager [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Terminating instance [ 746.122871] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "refresh_cache-8656c3b8-b7b1-442a-9719-1588ee83d19b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.123037] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquired lock "refresh_cache-8656c3b8-b7b1-442a-9719-1588ee83d19b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.123202] env[61867]: DEBUG nova.network.neutron [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 746.488022] env[61867]: DEBUG nova.compute.manager [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 746.507692] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance d0db58a5-d773-4e45-a238-07cfc156bdd2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 746.507806] env[61867]: DEBUG nova.network.neutron [-] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.510792] env[61867]: DEBUG nova.virt.hardware [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 746.511103] env[61867]: DEBUG nova.virt.hardware [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 746.511191] env[61867]: DEBUG nova.virt.hardware [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.511369] env[61867]: DEBUG nova.virt.hardware [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 746.511520] env[61867]: DEBUG nova.virt.hardware [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.511676] env[61867]: DEBUG nova.virt.hardware [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 746.511880] env[61867]: DEBUG nova.virt.hardware [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 746.512046] env[61867]: DEBUG nova.virt.hardware [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 746.512215] env[61867]: DEBUG nova.virt.hardware [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 746.512376] env[61867]: DEBUG nova.virt.hardware [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 746.512546] env[61867]: DEBUG nova.virt.hardware [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.514708] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2723413f-c67c-4dc3-a81c-950fa656dba3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.526935] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8cc6f8f-ec08-4dbd-abfe-808e20b6e692 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.613055] env[61867]: ERROR nova.compute.manager [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3b1112dd-2d7d-453e-ad9e-677480ae22c4, please check neutron logs for more information. [ 746.613055] env[61867]: ERROR nova.compute.manager Traceback (most recent call last): [ 746.613055] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.613055] env[61867]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 746.613055] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.613055] env[61867]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 746.613055] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.613055] env[61867]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 746.613055] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.613055] env[61867]: ERROR nova.compute.manager self.force_reraise() [ 746.613055] env[61867]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.613055] env[61867]: ERROR nova.compute.manager raise self.value [ 746.613055] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.613055] env[61867]: ERROR nova.compute.manager updated_port = self._update_port( [ 746.613055] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.613055] env[61867]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 746.613617] env[61867]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.613617] env[61867]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 746.613617] env[61867]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3b1112dd-2d7d-453e-ad9e-677480ae22c4, please check neutron logs for more information. [ 746.613617] env[61867]: ERROR nova.compute.manager [ 746.613617] env[61867]: Traceback (most recent call last): [ 746.613617] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 746.613617] env[61867]: listener.cb(fileno) [ 746.613617] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.613617] env[61867]: result = function(*args, **kwargs) [ 746.613617] env[61867]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 746.613617] env[61867]: return func(*args, **kwargs) [ 746.613617] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.613617] env[61867]: raise e [ 746.613617] env[61867]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.613617] env[61867]: nwinfo = self.network_api.allocate_for_instance( [ 746.613617] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.613617] env[61867]: created_port_ids = self._update_ports_for_instance( [ 746.613617] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.613617] env[61867]: with excutils.save_and_reraise_exception(): [ 746.613617] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.613617] env[61867]: self.force_reraise() [ 746.613617] env[61867]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.613617] env[61867]: raise self.value [ 746.613617] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.613617] env[61867]: updated_port = self._update_port( [ 746.613617] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.613617] env[61867]: _ensure_no_port_binding_failure(port) [ 746.613617] env[61867]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.613617] env[61867]: raise exception.PortBindingFailed(port_id=port['id']) [ 746.614660] env[61867]: nova.exception.PortBindingFailed: Binding failed for port 3b1112dd-2d7d-453e-ad9e-677480ae22c4, please check neutron logs for more information. [ 746.614660] env[61867]: Removing descriptor: 21 [ 746.614660] env[61867]: ERROR nova.compute.manager [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3b1112dd-2d7d-453e-ad9e-677480ae22c4, please check neutron logs for more information. [ 746.614660] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Traceback (most recent call last): [ 746.614660] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 746.614660] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] yield resources [ 746.614660] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 746.614660] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] self.driver.spawn(context, instance, image_meta, [ 746.614660] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 746.614660] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 746.614660] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 746.614660] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] vm_ref = self.build_virtual_machine(instance, [ 746.615086] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 746.615086] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] vif_infos = vmwarevif.get_vif_info(self._session, [ 746.615086] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 746.615086] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] for vif in network_info: [ 746.615086] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 746.615086] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] return self._sync_wrapper(fn, *args, **kwargs) [ 746.615086] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 746.615086] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] self.wait() [ 746.615086] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 746.615086] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] self[:] = self._gt.wait() [ 746.615086] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 746.615086] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] return self._exit_event.wait() [ 746.615086] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 746.615528] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] result = hub.switch() [ 746.615528] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 746.615528] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] return self.greenlet.switch() [ 746.615528] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.615528] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] result = function(*args, **kwargs) [ 746.615528] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 746.615528] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] return func(*args, **kwargs) [ 746.615528] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.615528] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] raise e [ 746.615528] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.615528] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] nwinfo = self.network_api.allocate_for_instance( [ 746.615528] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.615528] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] created_port_ids = self._update_ports_for_instance( [ 746.615956] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.615956] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] with excutils.save_and_reraise_exception(): [ 746.615956] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.615956] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] self.force_reraise() [ 746.615956] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.615956] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] raise self.value [ 746.615956] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.615956] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] updated_port = self._update_port( [ 746.615956] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.615956] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] _ensure_no_port_binding_failure(port) [ 746.615956] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.615956] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] raise exception.PortBindingFailed(port_id=port['id']) [ 746.616370] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] nova.exception.PortBindingFailed: Binding failed for port 3b1112dd-2d7d-453e-ad9e-677480ae22c4, please check neutron logs for more information. [ 746.616370] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] [ 746.616370] env[61867]: INFO nova.compute.manager [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Terminating instance [ 746.616482] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "refresh_cache-311a8d4f-68dc-4e0c-acdb-d3929afb9a1b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.616573] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "refresh_cache-311a8d4f-68dc-4e0c-acdb-d3929afb9a1b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.616735] env[61867]: DEBUG nova.network.neutron [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 746.639063] env[61867]: DEBUG nova.network.neutron [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.690200] env[61867]: DEBUG nova.network.neutron [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.818118] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.818299] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.018869] env[61867]: INFO nova.compute.manager [-] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Took 1.04 seconds to deallocate network for instance. [ 747.019677] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance b5a6ff7c-2b48-4f82-ba53-ec1977736c8f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 747.023551] env[61867]: DEBUG nova.compute.claims [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 747.024219] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.138029] env[61867]: DEBUG nova.network.neutron [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.193196] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Releasing lock "refresh_cache-8656c3b8-b7b1-442a-9719-1588ee83d19b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.193658] env[61867]: DEBUG nova.compute.manager [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 747.193995] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 747.195147] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77b86262-3f1c-45f5-a535-78840dd09273 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.202519] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 747.202860] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b226ecf2-ce37-43f8-80e0-930b67f36886 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.208222] env[61867]: DEBUG oslo_vmware.api [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 747.208222] env[61867]: value = "task-1276389" [ 747.208222] env[61867]: _type = "Task" [ 747.208222] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.215646] env[61867]: DEBUG oslo_vmware.api [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276389, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.216413] env[61867]: DEBUG nova.network.neutron [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.524388] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 81ada1e9-b100-44fd-8119-7ef8695e601f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 747.717744] env[61867]: DEBUG oslo_vmware.api [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276389, 'name': PowerOffVM_Task, 'duration_secs': 0.199347} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.718026] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 747.718205] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 747.718749] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "refresh_cache-311a8d4f-68dc-4e0c-acdb-d3929afb9a1b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.719163] env[61867]: DEBUG nova.compute.manager [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 747.719364] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 747.719613] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7218838f-db03-4dee-8296-4704ce10f9be {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.721423] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3bad3bf3-c2d1-45e7-a7ae-bb9ca8952bf3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.730414] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f79b64-4329-444e-801f-7690112984a1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.749114] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 747.749162] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 747.749334] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Deleting the datastore file [datastore2] 8656c3b8-b7b1-442a-9719-1588ee83d19b {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 747.753294] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d1716e0f-e8f3-42a6-8b0f-ad24be3814d6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.755309] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b could not be found. [ 747.755517] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 747.755697] env[61867]: INFO nova.compute.manager [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 747.755928] env[61867]: DEBUG oslo.service.loopingcall [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 747.756251] env[61867]: DEBUG nova.compute.manager [-] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 747.756360] env[61867]: DEBUG nova.network.neutron [-] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 747.760662] env[61867]: DEBUG nova.compute.manager [req-0c1fb1ba-dd2c-4ed2-b504-cc652c6bb678 req-13dba9bd-df65-4f9b-b04e-5a2e26393c3c service nova] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Received event network-changed-3b1112dd-2d7d-453e-ad9e-677480ae22c4 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 747.760662] env[61867]: DEBUG nova.compute.manager [req-0c1fb1ba-dd2c-4ed2-b504-cc652c6bb678 req-13dba9bd-df65-4f9b-b04e-5a2e26393c3c service nova] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Refreshing instance network info cache due to event network-changed-3b1112dd-2d7d-453e-ad9e-677480ae22c4. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 747.760662] env[61867]: DEBUG oslo_concurrency.lockutils [req-0c1fb1ba-dd2c-4ed2-b504-cc652c6bb678 req-13dba9bd-df65-4f9b-b04e-5a2e26393c3c service nova] Acquiring lock "refresh_cache-311a8d4f-68dc-4e0c-acdb-d3929afb9a1b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.760662] env[61867]: DEBUG oslo_concurrency.lockutils [req-0c1fb1ba-dd2c-4ed2-b504-cc652c6bb678 req-13dba9bd-df65-4f9b-b04e-5a2e26393c3c service nova] Acquired lock "refresh_cache-311a8d4f-68dc-4e0c-acdb-d3929afb9a1b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.760662] env[61867]: DEBUG nova.network.neutron [req-0c1fb1ba-dd2c-4ed2-b504-cc652c6bb678 req-13dba9bd-df65-4f9b-b04e-5a2e26393c3c service nova] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Refreshing network info cache for port 3b1112dd-2d7d-453e-ad9e-677480ae22c4 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 747.763680] env[61867]: DEBUG oslo_vmware.api [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 747.763680] env[61867]: value = "task-1276391" [ 747.763680] env[61867]: _type = "Task" [ 747.763680] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.772601] env[61867]: DEBUG oslo_vmware.api [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276391, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.775732] env[61867]: DEBUG nova.network.neutron [-] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.027395] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance c7203cea-2c2d-4a97-8ff4-3d33cd8443f7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 748.275778] env[61867]: DEBUG oslo_vmware.api [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276391, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09385} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.276043] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 748.276225] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 748.276394] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 748.276564] env[61867]: INFO nova.compute.manager [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Took 1.08 seconds to destroy the instance on the hypervisor. [ 748.276792] env[61867]: DEBUG oslo.service.loopingcall [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 748.276974] env[61867]: DEBUG nova.compute.manager [-] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.277075] env[61867]: DEBUG nova.network.neutron [-] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 748.278695] env[61867]: DEBUG nova.network.neutron [-] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.281838] env[61867]: DEBUG nova.network.neutron [req-0c1fb1ba-dd2c-4ed2-b504-cc652c6bb678 req-13dba9bd-df65-4f9b-b04e-5a2e26393c3c service nova] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.298174] env[61867]: DEBUG nova.network.neutron [-] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.363116] env[61867]: DEBUG nova.network.neutron [req-0c1fb1ba-dd2c-4ed2-b504-cc652c6bb678 req-13dba9bd-df65-4f9b-b04e-5a2e26393c3c service nova] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.530715] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance f0b6273f-4177-44a3-8dd1-f65faeb07539 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 748.781476] env[61867]: INFO nova.compute.manager [-] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Took 1.02 seconds to deallocate network for instance. [ 748.783886] env[61867]: DEBUG nova.compute.claims [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Aborting claim: {{(pid=61867) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 748.784288] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.801032] env[61867]: DEBUG nova.network.neutron [-] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.865619] env[61867]: DEBUG oslo_concurrency.lockutils [req-0c1fb1ba-dd2c-4ed2-b504-cc652c6bb678 req-13dba9bd-df65-4f9b-b04e-5a2e26393c3c service nova] Releasing lock "refresh_cache-311a8d4f-68dc-4e0c-acdb-d3929afb9a1b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.865619] env[61867]: DEBUG nova.compute.manager [req-0c1fb1ba-dd2c-4ed2-b504-cc652c6bb678 req-13dba9bd-df65-4f9b-b04e-5a2e26393c3c service nova] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Received event network-vif-deleted-3b1112dd-2d7d-453e-ad9e-677480ae22c4 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.034055] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 905ee661-d5bd-4121-9bd0-fe68bdabd2fd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 749.302851] env[61867]: INFO nova.compute.manager [-] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Took 1.03 seconds to deallocate network for instance. [ 749.537030] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 3022ea79-7c27-40a4-80d6-1a86b354917a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 749.809999] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.040813] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance be825f62-cb56-4647-be1c-1605536fdc89 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 750.543870] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 80a703b3-d692-4023-a73b-dba980a94dff has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 751.047242] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 1df8427c-e75d-4b60-a92a-b5ba76b67081 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 751.551037] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 81e70c36-04ea-450c-9383-53ef069d1c46 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 752.053713] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 4653ce96-257e-46d8-8c3d-85c03380213c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 752.558032] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 66518b2a-0242-438b-ba9f-d57c07a1165c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 753.060925] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance abb41c0c-6d0d-4147-a4af-554ab7d9e921 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 753.563424] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 6e41989e-b8fa-4009-af1e-1ce859b329a1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 754.066861] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 8a83f4f2-58eb-473e-9b1e-32ce633554f9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 754.067179] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 754.067313] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 754.321239] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1579421b-b5a1-4e6d-8d3a-417635cc23bb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.328466] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1351a4f-2424-47c9-8dd5-4177b3424dcd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.356708] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9aa7356-5fe7-405f-adc2-9f59fdc36296 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.363197] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c1e662-d753-4193-96a5-12971d9074ca {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.375375] env[61867]: DEBUG nova.compute.provider_tree [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.878291] env[61867]: DEBUG nova.scheduler.client.report [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 755.383882] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61867) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 755.384231] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.931s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.384309] env[61867]: DEBUG oslo_concurrency.lockutils [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.781s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.385922] env[61867]: INFO nova.compute.claims [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 756.628216] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8614394-6401-4692-a786-db88584026ce {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.635383] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911fa096-8e08-4fe0-bdde-23beea57c8fb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.664731] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34d840c-d012-42e7-aaf1-15becb2e862a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.671502] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3286fcc4-9a77-4b01-bb00-6a39b8ee4d35 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.684491] env[61867]: DEBUG nova.compute.provider_tree [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.187744] env[61867]: DEBUG nova.scheduler.client.report [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 757.694505] env[61867]: DEBUG oslo_concurrency.lockutils [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.694505] env[61867]: DEBUG nova.compute.manager [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 757.697091] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.899s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.699081] env[61867]: INFO nova.compute.claims [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 757.957153] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 757.957341] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Cleaning up deleted instances {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 758.202892] env[61867]: DEBUG nova.compute.utils [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 758.206110] env[61867]: DEBUG nova.compute.manager [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 758.206280] env[61867]: DEBUG nova.network.neutron [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 758.257073] env[61867]: DEBUG nova.policy [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a53cf3a1379847a1bc70f60d739f7aa0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcacc42585c145dcba63134936c030a6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 758.461277] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] There are 3 instances to clean {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 758.461580] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 37ed9e18-8dba-459d-bc67-e3ce3f9cdaea] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 758.520776] env[61867]: DEBUG nova.network.neutron [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Successfully created port: bb4a3c79-631f-4359-927a-a2a7b2624aa9 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 758.712129] env[61867]: DEBUG nova.compute.manager [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 758.964519] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 76d3a228-e314-42b1-9eec-43a32fbd8035] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 758.975602] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc464f46-1c02-4a1b-92f4-847f2fd1aaae {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.982641] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-710b6a9f-8f19-492f-9adf-b4ce8124a7d3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.015032] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d4115a-ddc9-4baf-9631-8fbca0be8271 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.022028] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1521562-f286-4c46-a861-546db0a91b34 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.034911] env[61867]: DEBUG nova.compute.provider_tree [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.468292] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 318ecaf7-7e12-4d42-9b2f-3f3ab2577ffa] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 759.537858] env[61867]: DEBUG nova.scheduler.client.report [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.722409] env[61867]: DEBUG nova.compute.manager [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 759.746222] env[61867]: DEBUG nova.virt.hardware [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 759.746488] env[61867]: DEBUG nova.virt.hardware [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 759.746642] env[61867]: DEBUG nova.virt.hardware [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 759.746817] env[61867]: DEBUG nova.virt.hardware [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 759.746955] env[61867]: DEBUG nova.virt.hardware [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 759.747110] env[61867]: DEBUG nova.virt.hardware [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 759.747307] env[61867]: DEBUG nova.virt.hardware [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 759.747465] env[61867]: DEBUG nova.virt.hardware [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 759.747616] env[61867]: DEBUG nova.virt.hardware [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 759.747774] env[61867]: DEBUG nova.virt.hardware [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 759.747939] env[61867]: DEBUG nova.virt.hardware [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 759.748780] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba57ba0b-b1a1-4413-a760-1e23beb1c012 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.756383] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197e0d59-7965-4443-9e6d-4793cfff9d70 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.932583] env[61867]: DEBUG nova.compute.manager [req-3025f470-f2bd-4cde-be0b-867d267ac2a7 req-d34fa209-37c0-42e8-92aa-c9e2fff9942d service nova] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Received event network-vif-plugged-bb4a3c79-631f-4359-927a-a2a7b2624aa9 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.932938] env[61867]: DEBUG oslo_concurrency.lockutils [req-3025f470-f2bd-4cde-be0b-867d267ac2a7 req-d34fa209-37c0-42e8-92aa-c9e2fff9942d service nova] Acquiring lock "7c0badee-0e2b-4a13-89b1-e57df5c2c9ef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.933661] env[61867]: DEBUG oslo_concurrency.lockutils [req-3025f470-f2bd-4cde-be0b-867d267ac2a7 req-d34fa209-37c0-42e8-92aa-c9e2fff9942d service nova] Lock "7c0badee-0e2b-4a13-89b1-e57df5c2c9ef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.933996] env[61867]: DEBUG oslo_concurrency.lockutils [req-3025f470-f2bd-4cde-be0b-867d267ac2a7 req-d34fa209-37c0-42e8-92aa-c9e2fff9942d service nova] Lock "7c0badee-0e2b-4a13-89b1-e57df5c2c9ef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.934398] env[61867]: DEBUG nova.compute.manager [req-3025f470-f2bd-4cde-be0b-867d267ac2a7 req-d34fa209-37c0-42e8-92aa-c9e2fff9942d service nova] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] No waiting events found dispatching network-vif-plugged-bb4a3c79-631f-4359-927a-a2a7b2624aa9 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 759.934761] env[61867]: WARNING nova.compute.manager [req-3025f470-f2bd-4cde-be0b-867d267ac2a7 req-d34fa209-37c0-42e8-92aa-c9e2fff9942d service nova] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Received unexpected event network-vif-plugged-bb4a3c79-631f-4359-927a-a2a7b2624aa9 for instance with vm_state building and task_state spawning. [ 759.972497] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 759.973770] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Cleaning up deleted instances with incomplete migration {{(pid=61867) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 760.032628] env[61867]: DEBUG nova.network.neutron [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Successfully updated port: bb4a3c79-631f-4359-927a-a2a7b2624aa9 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 760.042190] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.042700] env[61867]: DEBUG nova.compute.manager [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 760.045095] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.787s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.046560] env[61867]: INFO nova.compute.claims [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 760.478842] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.535693] env[61867]: DEBUG oslo_concurrency.lockutils [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Acquiring lock "refresh_cache-7c0badee-0e2b-4a13-89b1-e57df5c2c9ef" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.535861] env[61867]: DEBUG oslo_concurrency.lockutils [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Acquired lock "refresh_cache-7c0badee-0e2b-4a13-89b1-e57df5c2c9ef" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.536024] env[61867]: DEBUG nova.network.neutron [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 760.551394] env[61867]: DEBUG nova.compute.utils [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 760.554505] env[61867]: DEBUG nova.compute.manager [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Not allocating networking since 'none' was specified. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 761.055379] env[61867]: DEBUG nova.compute.manager [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 761.074310] env[61867]: DEBUG nova.network.neutron [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.265427] env[61867]: DEBUG nova.network.neutron [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Updating instance_info_cache with network_info: [{"id": "bb4a3c79-631f-4359-927a-a2a7b2624aa9", "address": "fa:16:3e:fb:a9:88", "network": {"id": "cae226ee-fbc8-4f23-81bd-264c168ab311", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1212001364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcacc42585c145dcba63134936c030a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb4a3c79-63", "ovs_interfaceid": "bb4a3c79-631f-4359-927a-a2a7b2624aa9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.327567] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f85cbfff-5b5e-4d12-8dae-32953f8a4794 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.334829] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f8e046-c1ba-4d82-b48c-2f54021f75bf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.364943] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4273e3a-7bfc-4019-a0f8-1e1d9a5c9e39 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.372073] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8e6f1d4-026f-469d-89a8-c054635d7405 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.385010] env[61867]: DEBUG nova.compute.provider_tree [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.767853] env[61867]: DEBUG oslo_concurrency.lockutils [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Releasing lock "refresh_cache-7c0badee-0e2b-4a13-89b1-e57df5c2c9ef" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.768281] env[61867]: DEBUG nova.compute.manager [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Instance network_info: |[{"id": "bb4a3c79-631f-4359-927a-a2a7b2624aa9", "address": "fa:16:3e:fb:a9:88", "network": {"id": "cae226ee-fbc8-4f23-81bd-264c168ab311", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1212001364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcacc42585c145dcba63134936c030a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb4a3c79-63", "ovs_interfaceid": "bb4a3c79-631f-4359-927a-a2a7b2624aa9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 761.768795] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:a9:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4db2961d-273d-4634-9d06-a94fa9d384fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bb4a3c79-631f-4359-927a-a2a7b2624aa9', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 761.776482] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Creating folder: Project (fcacc42585c145dcba63134936c030a6). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 761.776777] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-deceab6e-dd4c-4e78-8bbd-5d8adcad8c35 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.787144] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Created folder: Project (fcacc42585c145dcba63134936c030a6) in parent group-v274258. [ 761.787321] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Creating folder: Instances. Parent ref: group-v274284. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 761.787535] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d700f22b-396d-4879-bd2b-b0baec93edf7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.795718] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Created folder: Instances in parent group-v274284. [ 761.795931] env[61867]: DEBUG oslo.service.loopingcall [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.796112] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 761.796291] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4dce36b3-9da6-435b-a2b9-a30358bc5c56 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.813764] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 761.813764] env[61867]: value = "task-1276394" [ 761.813764] env[61867]: _type = "Task" [ 761.813764] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.824476] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276394, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.888478] env[61867]: DEBUG nova.scheduler.client.report [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.959467] env[61867]: DEBUG nova.compute.manager [req-6e3c8539-e797-49f7-b1ec-e66334a89429 req-ecebe085-2ba0-46bd-be7c-de675b00ec56 service nova] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Received event network-changed-bb4a3c79-631f-4359-927a-a2a7b2624aa9 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.960756] env[61867]: DEBUG nova.compute.manager [req-6e3c8539-e797-49f7-b1ec-e66334a89429 req-ecebe085-2ba0-46bd-be7c-de675b00ec56 service nova] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Refreshing instance network info cache due to event network-changed-bb4a3c79-631f-4359-927a-a2a7b2624aa9. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 761.960756] env[61867]: DEBUG oslo_concurrency.lockutils [req-6e3c8539-e797-49f7-b1ec-e66334a89429 req-ecebe085-2ba0-46bd-be7c-de675b00ec56 service nova] Acquiring lock "refresh_cache-7c0badee-0e2b-4a13-89b1-e57df5c2c9ef" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.960756] env[61867]: DEBUG oslo_concurrency.lockutils [req-6e3c8539-e797-49f7-b1ec-e66334a89429 req-ecebe085-2ba0-46bd-be7c-de675b00ec56 service nova] Acquired lock "refresh_cache-7c0badee-0e2b-4a13-89b1-e57df5c2c9ef" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.960923] env[61867]: DEBUG nova.network.neutron [req-6e3c8539-e797-49f7-b1ec-e66334a89429 req-ecebe085-2ba0-46bd-be7c-de675b00ec56 service nova] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Refreshing network info cache for port bb4a3c79-631f-4359-927a-a2a7b2624aa9 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 761.975723] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.975920] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.976080] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Starting heal instance info cache {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 761.976201] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Rebuilding the list of instances to heal {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 762.067491] env[61867]: DEBUG nova.compute.manager [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 762.092452] env[61867]: DEBUG nova.virt.hardware [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 762.092705] env[61867]: DEBUG nova.virt.hardware [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 762.092864] env[61867]: DEBUG nova.virt.hardware [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 762.093055] env[61867]: DEBUG nova.virt.hardware [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 762.093202] env[61867]: DEBUG nova.virt.hardware [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 762.093348] env[61867]: DEBUG nova.virt.hardware [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 762.093553] env[61867]: DEBUG nova.virt.hardware [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 762.093733] env[61867]: DEBUG nova.virt.hardware [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 762.093925] env[61867]: DEBUG nova.virt.hardware [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 762.094104] env[61867]: DEBUG nova.virt.hardware [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 762.094280] env[61867]: DEBUG nova.virt.hardware [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 762.095463] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e32b57d3-897f-452a-8aab-90b53d00c0fa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.103264] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30776ad1-b3d2-4c64-97e5-41b8463b7967 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.117442] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Instance VIF info [] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 762.122890] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Creating folder: Project (5ab0d8a0642643ea909b2d4faacaecfd). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 762.123186] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5548a9b0-330f-482d-87bf-82ed26bb1970 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.132116] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Created folder: Project (5ab0d8a0642643ea909b2d4faacaecfd) in parent group-v274258. [ 762.132301] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Creating folder: Instances. Parent ref: group-v274287. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 762.132512] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-548bbdef-c12b-475b-879d-a3cb80c2566f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.140765] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Created folder: Instances in parent group-v274287. [ 762.140953] env[61867]: DEBUG oslo.service.loopingcall [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.141148] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 762.141336] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-00e32698-3328-4a83-b8ab-bf6d496ac54b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.156758] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 762.156758] env[61867]: value = "task-1276397" [ 762.156758] env[61867]: _type = "Task" [ 762.156758] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.164091] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276397, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.323999] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276394, 'name': CreateVM_Task, 'duration_secs': 0.281531} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.324199] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 762.331311] env[61867]: DEBUG oslo_concurrency.lockutils [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.331482] env[61867]: DEBUG oslo_concurrency.lockutils [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.331796] env[61867]: DEBUG oslo_concurrency.lockutils [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 762.332051] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0215d3c-6b56-4e17-938e-381c7217896d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.336528] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Waiting for the task: (returnval){ [ 762.336528] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522b52be-6311-3e9b-709e-2d0a216d5d0b" [ 762.336528] env[61867]: _type = "Task" [ 762.336528] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.344630] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522b52be-6311-3e9b-709e-2d0a216d5d0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.393738] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.349s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.394332] env[61867]: DEBUG nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 762.397518] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.862s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.398508] env[61867]: INFO nova.compute.claims [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.482211] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.482211] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.482211] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.482211] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.482211] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.482211] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.482513] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.514229] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "refresh_cache-b1ec32c5-0642-4dca-ad43-05a5172d04d7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.514381] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquired lock "refresh_cache-b1ec32c5-0642-4dca-ad43-05a5172d04d7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.514535] env[61867]: DEBUG nova.network.neutron [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Forcefully refreshing network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 762.514698] env[61867]: DEBUG nova.objects.instance [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lazy-loading 'info_cache' on Instance uuid b1ec32c5-0642-4dca-ad43-05a5172d04d7 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 762.667181] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276397, 'name': CreateVM_Task, 'duration_secs': 0.242751} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.667352] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 762.667759] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.674238] env[61867]: DEBUG nova.network.neutron [req-6e3c8539-e797-49f7-b1ec-e66334a89429 req-ecebe085-2ba0-46bd-be7c-de675b00ec56 service nova] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Updated VIF entry in instance network info cache for port bb4a3c79-631f-4359-927a-a2a7b2624aa9. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 762.674621] env[61867]: DEBUG nova.network.neutron [req-6e3c8539-e797-49f7-b1ec-e66334a89429 req-ecebe085-2ba0-46bd-be7c-de675b00ec56 service nova] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Updating instance_info_cache with network_info: [{"id": "bb4a3c79-631f-4359-927a-a2a7b2624aa9", "address": "fa:16:3e:fb:a9:88", "network": {"id": "cae226ee-fbc8-4f23-81bd-264c168ab311", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1212001364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcacc42585c145dcba63134936c030a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb4a3c79-63", "ovs_interfaceid": "bb4a3c79-631f-4359-927a-a2a7b2624aa9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.846782] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522b52be-6311-3e9b-709e-2d0a216d5d0b, 'name': SearchDatastore_Task, 'duration_secs': 0.009493} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.848056] env[61867]: DEBUG oslo_concurrency.lockutils [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.848056] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 762.848056] env[61867]: DEBUG oslo_concurrency.lockutils [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.848056] env[61867]: DEBUG oslo_concurrency.lockutils [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.848293] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 762.848429] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.848722] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 762.848934] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-438d414f-1f8d-485d-b437-ab3301cf4015 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.850683] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc82419a-968b-48ae-b568-7bedb664a6b3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.855833] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Waiting for the task: (returnval){ [ 762.855833] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5224ca92-a0d3-91af-c36d-cb83c4414abb" [ 762.855833] env[61867]: _type = "Task" [ 762.855833] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.859342] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 762.859512] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 762.860494] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1894d88f-d12d-44ab-a6de-ea168df9532a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.865147] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5224ca92-a0d3-91af-c36d-cb83c4414abb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.867700] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Waiting for the task: (returnval){ [ 762.867700] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5219c974-d97c-10b0-fcd9-c158adbc2ee0" [ 762.867700] env[61867]: _type = "Task" [ 762.867700] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.874471] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5219c974-d97c-10b0-fcd9-c158adbc2ee0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.902734] env[61867]: DEBUG nova.compute.utils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 762.905739] env[61867]: DEBUG nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 762.905908] env[61867]: DEBUG nova.network.neutron [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 762.944440] env[61867]: DEBUG nova.policy [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9781b7a61f15483b8577f854266a876d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8451dfe23754279965eda6b2f2f37cf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 763.177613] env[61867]: DEBUG oslo_concurrency.lockutils [req-6e3c8539-e797-49f7-b1ec-e66334a89429 req-ecebe085-2ba0-46bd-be7c-de675b00ec56 service nova] Releasing lock "refresh_cache-7c0badee-0e2b-4a13-89b1-e57df5c2c9ef" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.190771] env[61867]: DEBUG nova.network.neutron [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Successfully created port: 5661376f-2454-4693-a106-69cb78133962 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 763.366293] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5224ca92-a0d3-91af-c36d-cb83c4414abb, 'name': SearchDatastore_Task, 'duration_secs': 0.020059} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.366622] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.366860] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 763.367077] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.376166] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5219c974-d97c-10b0-fcd9-c158adbc2ee0, 'name': SearchDatastore_Task, 'duration_secs': 0.009941} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.376887] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ca59699-eca6-4e50-9566-5b46b94bc166 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.381609] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Waiting for the task: (returnval){ [ 763.381609] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520c0b12-79a7-3a18-ee75-0f49edb2f8df" [ 763.381609] env[61867]: _type = "Task" [ 763.381609] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.388910] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520c0b12-79a7-3a18-ee75-0f49edb2f8df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.406542] env[61867]: DEBUG nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 763.536601] env[61867]: DEBUG nova.network.neutron [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.675692] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b649267f-876b-446b-a1b6-8457488f6c9c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.683213] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be7d8b6-f8c5-437e-a27c-d4d853bc3941 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.713237] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3857523f-f825-46a7-b4f6-14d8e729dbed {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.720845] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578ef35f-5e76-42e6-9f34-88033ab34bf1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.733755] env[61867]: DEBUG nova.compute.provider_tree [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.892970] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520c0b12-79a7-3a18-ee75-0f49edb2f8df, 'name': SearchDatastore_Task, 'duration_secs': 0.009208} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.893248] env[61867]: DEBUG oslo_concurrency.lockutils [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.893496] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef/7c0badee-0e2b-4a13-89b1-e57df5c2c9ef.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 763.893756] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.893936] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 763.894145] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-26ebe25b-7a0e-40c3-929b-626bed9e4871 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.895898] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea5c7c14-79ba-469f-b611-f96562a371d7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.902135] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Waiting for the task: (returnval){ [ 763.902135] env[61867]: value = "task-1276398" [ 763.902135] env[61867]: _type = "Task" [ 763.902135] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.905561] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 763.905729] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 763.906650] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03dbe749-26bd-4ee7-acdd-00f493624295 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.911532] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': task-1276398, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.915956] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Waiting for the task: (returnval){ [ 763.915956] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b2da14-fb65-0937-6021-b30112f9e13a" [ 763.915956] env[61867]: _type = "Task" [ 763.915956] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.923606] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b2da14-fb65-0937-6021-b30112f9e13a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.098435] env[61867]: DEBUG nova.network.neutron [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.237070] env[61867]: DEBUG nova.scheduler.client.report [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.413067] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': task-1276398, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.437661} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.413067] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef/7c0badee-0e2b-4a13-89b1-e57df5c2c9ef.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 764.413067] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 764.413067] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-24f79ba5-aac4-4f0c-86cb-3d28ecf0bc89 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.417909] env[61867]: DEBUG nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 764.424321] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Waiting for the task: (returnval){ [ 764.424321] env[61867]: value = "task-1276399" [ 764.424321] env[61867]: _type = "Task" [ 764.424321] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.431022] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b2da14-fb65-0937-6021-b30112f9e13a, 'name': SearchDatastore_Task, 'duration_secs': 0.008596} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.431719] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6358ee38-4a29-4058-9bf0-1fd6aa9fa7e9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.436700] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': task-1276399, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.441331] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Waiting for the task: (returnval){ [ 764.441331] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52bdb077-f252-910d-db9a-da29c3e0c163" [ 764.441331] env[61867]: _type = "Task" [ 764.441331] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.446348] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 764.446732] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 764.446732] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 764.446900] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 764.447057] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 764.447203] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 764.447415] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 764.447587] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 764.447750] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 764.447910] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 764.448114] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 764.448864] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ab9363-364a-4786-83b0-f774c2323dec {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.456196] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52bdb077-f252-910d-db9a-da29c3e0c163, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.460592] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ef2687-b4ea-4df5-8d89-b7d45087f778 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.595765] env[61867]: DEBUG nova.compute.manager [req-10f8f71f-db95-4e35-89be-efef6f5c31c9 req-11d3d251-a885-444a-b467-dd3b77549147 service nova] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Received event network-vif-plugged-5661376f-2454-4693-a106-69cb78133962 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.595984] env[61867]: DEBUG oslo_concurrency.lockutils [req-10f8f71f-db95-4e35-89be-efef6f5c31c9 req-11d3d251-a885-444a-b467-dd3b77549147 service nova] Acquiring lock "b5a6ff7c-2b48-4f82-ba53-ec1977736c8f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.596215] env[61867]: DEBUG oslo_concurrency.lockutils [req-10f8f71f-db95-4e35-89be-efef6f5c31c9 req-11d3d251-a885-444a-b467-dd3b77549147 service nova] Lock "b5a6ff7c-2b48-4f82-ba53-ec1977736c8f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.596383] env[61867]: DEBUG oslo_concurrency.lockutils [req-10f8f71f-db95-4e35-89be-efef6f5c31c9 req-11d3d251-a885-444a-b467-dd3b77549147 service nova] Lock "b5a6ff7c-2b48-4f82-ba53-ec1977736c8f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.596548] env[61867]: DEBUG nova.compute.manager [req-10f8f71f-db95-4e35-89be-efef6f5c31c9 req-11d3d251-a885-444a-b467-dd3b77549147 service nova] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] No waiting events found dispatching network-vif-plugged-5661376f-2454-4693-a106-69cb78133962 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 764.596709] env[61867]: WARNING nova.compute.manager [req-10f8f71f-db95-4e35-89be-efef6f5c31c9 req-11d3d251-a885-444a-b467-dd3b77549147 service nova] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Received unexpected event network-vif-plugged-5661376f-2454-4693-a106-69cb78133962 for instance with vm_state building and task_state spawning. [ 764.600996] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Releasing lock "refresh_cache-b1ec32c5-0642-4dca-ad43-05a5172d04d7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.601190] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Updated the network info_cache for instance {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 764.601373] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.601527] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.601669] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.601814] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.601952] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.602106] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.602232] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61867) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 764.602373] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.679677] env[61867]: DEBUG nova.network.neutron [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Successfully updated port: 5661376f-2454-4693-a106-69cb78133962 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 764.742710] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.743271] env[61867]: DEBUG nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 764.747212] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.136s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.933301] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': task-1276399, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.188895} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.934019] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 764.934560] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c36a1a-1a11-4abf-9dd3-40a11a0124a8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.955954] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef/7c0badee-0e2b-4a13-89b1-e57df5c2c9ef.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 764.958881] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0906f9f0-00dd-4270-b01b-56e225b06908 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.977349] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52bdb077-f252-910d-db9a-da29c3e0c163, 'name': SearchDatastore_Task, 'duration_secs': 0.015761} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.978490] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.978763] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] d0db58a5-d773-4e45-a238-07cfc156bdd2/d0db58a5-d773-4e45-a238-07cfc156bdd2.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 764.979076] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Waiting for the task: (returnval){ [ 764.979076] env[61867]: value = "task-1276400" [ 764.979076] env[61867]: _type = "Task" [ 764.979076] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.979272] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f888d571-4872-4f14-8f20-0112a370428b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.988569] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': task-1276400, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.989741] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Waiting for the task: (returnval){ [ 764.989741] env[61867]: value = "task-1276401" [ 764.989741] env[61867]: _type = "Task" [ 764.989741] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.996775] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': task-1276401, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.105709] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.182197] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "refresh_cache-b5a6ff7c-2b48-4f82-ba53-ec1977736c8f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.182365] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquired lock "refresh_cache-b5a6ff7c-2b48-4f82-ba53-ec1977736c8f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.182525] env[61867]: DEBUG nova.network.neutron [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 765.253383] env[61867]: DEBUG nova.compute.utils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.259552] env[61867]: DEBUG nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 765.260325] env[61867]: DEBUG nova.network.neutron [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 765.325678] env[61867]: DEBUG nova.policy [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9781b7a61f15483b8577f854266a876d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8451dfe23754279965eda6b2f2f37cf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 765.492562] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': task-1276400, 'name': ReconfigVM_Task, 'duration_secs': 0.325067} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.495481] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef/7c0badee-0e2b-4a13-89b1-e57df5c2c9ef.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 765.498125] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2bbb1684-4358-4f18-b401-41737de1b852 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.505046] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': task-1276401, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450422} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.505860] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] d0db58a5-d773-4e45-a238-07cfc156bdd2/d0db58a5-d773-4e45-a238-07cfc156bdd2.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 765.506041] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 765.506339] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Waiting for the task: (returnval){ [ 765.506339] env[61867]: value = "task-1276402" [ 765.506339] env[61867]: _type = "Task" [ 765.506339] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.506538] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3de15254-86af-467f-a233-5b96b49b8c80 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.518062] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': task-1276402, 'name': Rename_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.519111] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Waiting for the task: (returnval){ [ 765.519111] env[61867]: value = "task-1276403" [ 765.519111] env[61867]: _type = "Task" [ 765.519111] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.527939] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': task-1276403, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.584892] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a5fb6c7-400c-4eb3-926c-f024ef171ebf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.593690] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f9ece5c-ff91-475f-816d-9bc8edf04d09 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.623250] env[61867]: DEBUG nova.network.neutron [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Successfully created port: 471a814b-d572-4c6b-bdbf-9e02e2091d77 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 765.625628] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e354b1-a1ba-46ef-8d10-4befdbd2a050 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.633373] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf1b8084-a8af-4512-8207-e733a2d395d2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.646453] env[61867]: DEBUG nova.compute.provider_tree [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.729231] env[61867]: DEBUG nova.network.neutron [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.761850] env[61867]: DEBUG nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 765.952382] env[61867]: DEBUG nova.network.neutron [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Updating instance_info_cache with network_info: [{"id": "5661376f-2454-4693-a106-69cb78133962", "address": "fa:16:3e:7c:33:4d", "network": {"id": "998e68d3-8e78-4db8-8e53-859a2a9c494a", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1638393568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8451dfe23754279965eda6b2f2f37cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9297313e-7c50-4873-93d3-67284929163a", "external-id": "nsx-vlan-transportzone-620", "segmentation_id": 620, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5661376f-24", "ovs_interfaceid": "5661376f-2454-4693-a106-69cb78133962", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.017954] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': task-1276402, 'name': Rename_Task, 'duration_secs': 0.14973} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.018331] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 766.018596] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e4656507-47bf-4d87-95b1-bbf0de713fe3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.027985] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': task-1276403, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.149466} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.029146] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 766.029778] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Waiting for the task: (returnval){ [ 766.029778] env[61867]: value = "task-1276404" [ 766.029778] env[61867]: _type = "Task" [ 766.029778] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.030500] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0512542-2ecb-47b2-959b-42c6b5180915 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.041209] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': task-1276404, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.760795] env[61867]: DEBUG nova.scheduler.client.report [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 766.766574] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Releasing lock "refresh_cache-b5a6ff7c-2b48-4f82-ba53-ec1977736c8f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.766809] env[61867]: DEBUG nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Instance network_info: |[{"id": "5661376f-2454-4693-a106-69cb78133962", "address": "fa:16:3e:7c:33:4d", "network": {"id": "998e68d3-8e78-4db8-8e53-859a2a9c494a", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1638393568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8451dfe23754279965eda6b2f2f37cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9297313e-7c50-4873-93d3-67284929163a", "external-id": "nsx-vlan-transportzone-620", "segmentation_id": 620, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5661376f-24", "ovs_interfaceid": "5661376f-2454-4693-a106-69cb78133962", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 766.776590] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] d0db58a5-d773-4e45-a238-07cfc156bdd2/d0db58a5-d773-4e45-a238-07cfc156bdd2.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 766.777394] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7c:33:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9297313e-7c50-4873-93d3-67284929163a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5661376f-2454-4693-a106-69cb78133962', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 766.788023] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Creating folder: Project (b8451dfe23754279965eda6b2f2f37cf). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 766.788673] env[61867]: DEBUG nova.compute.manager [req-326dd822-5243-4c9a-86b6-ddd175306cb4 req-96791bc4-2d83-4e8e-972f-657bdfb4174b service nova] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Received event network-changed-5661376f-2454-4693-a106-69cb78133962 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 766.788845] env[61867]: DEBUG nova.compute.manager [req-326dd822-5243-4c9a-86b6-ddd175306cb4 req-96791bc4-2d83-4e8e-972f-657bdfb4174b service nova] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Refreshing instance network info cache due to event network-changed-5661376f-2454-4693-a106-69cb78133962. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 766.789064] env[61867]: DEBUG oslo_concurrency.lockutils [req-326dd822-5243-4c9a-86b6-ddd175306cb4 req-96791bc4-2d83-4e8e-972f-657bdfb4174b service nova] Acquiring lock "refresh_cache-b5a6ff7c-2b48-4f82-ba53-ec1977736c8f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.789200] env[61867]: DEBUG oslo_concurrency.lockutils [req-326dd822-5243-4c9a-86b6-ddd175306cb4 req-96791bc4-2d83-4e8e-972f-657bdfb4174b service nova] Acquired lock "refresh_cache-b5a6ff7c-2b48-4f82-ba53-ec1977736c8f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.789354] env[61867]: DEBUG nova.network.neutron [req-326dd822-5243-4c9a-86b6-ddd175306cb4 req-96791bc4-2d83-4e8e-972f-657bdfb4174b service nova] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Refreshing network info cache for port 5661376f-2454-4693-a106-69cb78133962 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 766.790591] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07c89ce0-f446-43b5-873a-fe01cace6fd7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.805468] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8b3ae446-168c-4477-97de-a8afa585b9e5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.815726] env[61867]: DEBUG oslo_vmware.api [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': task-1276404, 'name': PowerOnVM_Task, 'duration_secs': 0.470202} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.817689] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 766.817910] env[61867]: INFO nova.compute.manager [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Took 7.10 seconds to spawn the instance on the hypervisor. [ 766.818112] env[61867]: DEBUG nova.compute.manager [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 766.818445] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Waiting for the task: (returnval){ [ 766.818445] env[61867]: value = "task-1276405" [ 766.818445] env[61867]: _type = "Task" [ 766.818445] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.818864] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Created folder: Project (b8451dfe23754279965eda6b2f2f37cf) in parent group-v274258. [ 766.819048] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Creating folder: Instances. Parent ref: group-v274290. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 766.819782] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd2624e-dc3c-491a-bd8c-ec91365e997a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.822347] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-adc80523-73be-4aaf-bc21-52fc82892e4b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.831953] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': task-1276405, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.839569] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Created folder: Instances in parent group-v274290. [ 766.839836] env[61867]: DEBUG oslo.service.loopingcall [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 766.840020] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 766.840223] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a31daed8-f5c0-4bf8-8bf7-9a602ad5bde9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.859170] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 766.859170] env[61867]: value = "task-1276408" [ 766.859170] env[61867]: _type = "Task" [ 766.859170] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.867210] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276408, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.235442] env[61867]: DEBUG nova.network.neutron [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Successfully updated port: 471a814b-d572-4c6b-bdbf-9e02e2091d77 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 767.278018] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.531s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.278719] env[61867]: ERROR nova.compute.manager [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bfc894ba-9915-48bc-b7b2-83785909f7d4, please check neutron logs for more information. [ 767.278719] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Traceback (most recent call last): [ 767.278719] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 767.278719] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] self.driver.spawn(context, instance, image_meta, [ 767.278719] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 767.278719] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] self._vmops.spawn(context, instance, image_meta, injected_files, [ 767.278719] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 767.278719] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] vm_ref = self.build_virtual_machine(instance, [ 767.278719] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 767.278719] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] vif_infos = vmwarevif.get_vif_info(self._session, [ 767.278719] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 767.279077] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] for vif in network_info: [ 767.279077] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 767.279077] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] return self._sync_wrapper(fn, *args, **kwargs) [ 767.279077] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 767.279077] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] self.wait() [ 767.279077] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 767.279077] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] self[:] = self._gt.wait() [ 767.279077] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 767.279077] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] return self._exit_event.wait() [ 767.279077] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 767.279077] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] current.throw(*self._exc) [ 767.279077] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.279077] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] result = function(*args, **kwargs) [ 767.279445] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 767.279445] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] return func(*args, **kwargs) [ 767.279445] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 767.279445] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] raise e [ 767.279445] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.279445] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] nwinfo = self.network_api.allocate_for_instance( [ 767.279445] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.279445] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] created_port_ids = self._update_ports_for_instance( [ 767.279445] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.279445] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] with excutils.save_and_reraise_exception(): [ 767.279445] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.279445] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] self.force_reraise() [ 767.279445] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.279913] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] raise self.value [ 767.279913] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.279913] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] updated_port = self._update_port( [ 767.279913] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.279913] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] _ensure_no_port_binding_failure(port) [ 767.279913] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.279913] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] raise exception.PortBindingFailed(port_id=port['id']) [ 767.279913] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] nova.exception.PortBindingFailed: Binding failed for port bfc894ba-9915-48bc-b7b2-83785909f7d4, please check neutron logs for more information. [ 767.279913] env[61867]: ERROR nova.compute.manager [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] [ 767.279913] env[61867]: DEBUG nova.compute.utils [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Binding failed for port bfc894ba-9915-48bc-b7b2-83785909f7d4, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 767.281457] env[61867]: DEBUG nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 767.283607] env[61867]: DEBUG nova.compute.manager [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Build of instance 5e97f142-b5c0-4318-9a35-085234d1cb64 was re-scheduled: Binding failed for port bfc894ba-9915-48bc-b7b2-83785909f7d4, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 767.284060] env[61867]: DEBUG nova.compute.manager [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 767.284289] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Acquiring lock "refresh_cache-5e97f142-b5c0-4318-9a35-085234d1cb64" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.284434] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Acquired lock "refresh_cache-5e97f142-b5c0-4318-9a35-085234d1cb64" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.284630] env[61867]: DEBUG nova.network.neutron [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 767.285756] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.219s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.287194] env[61867]: INFO nova.compute.claims [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 767.309186] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 767.309444] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 767.309632] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 767.309891] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 767.310163] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 767.310231] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 767.310409] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 767.310563] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 767.310725] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 767.310884] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 767.311063] env[61867]: DEBUG nova.virt.hardware [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 767.312320] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49360f17-25f4-4506-9a15-d25e1eeb4f53 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.322396] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae1d27dd-a3ff-4135-b026-87c8987a8d8b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.345151] env[61867]: INFO nova.compute.manager [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Took 33.76 seconds to build instance. [ 767.351703] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': task-1276405, 'name': ReconfigVM_Task, 'duration_secs': 0.316142} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.352175] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Reconfigured VM instance instance-00000035 to attach disk [datastore2] d0db58a5-d773-4e45-a238-07cfc156bdd2/d0db58a5-d773-4e45-a238-07cfc156bdd2.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 767.353408] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-15c32fcf-24dd-4629-91c4-ddcad2ff4a40 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.359166] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Waiting for the task: (returnval){ [ 767.359166] env[61867]: value = "task-1276409" [ 767.359166] env[61867]: _type = "Task" [ 767.359166] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.372459] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': task-1276409, 'name': Rename_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.375158] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276408, 'name': CreateVM_Task, 'duration_secs': 0.385161} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.375311] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 767.375951] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.376124] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.376433] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 767.376667] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c36fcd3a-647e-4ee8-830c-7967d0c5c2c5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.380862] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 767.380862] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525e54fb-03e4-86cd-a1af-b52de615baaa" [ 767.380862] env[61867]: _type = "Task" [ 767.380862] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.389350] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525e54fb-03e4-86cd-a1af-b52de615baaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.664644] env[61867]: DEBUG nova.network.neutron [req-326dd822-5243-4c9a-86b6-ddd175306cb4 req-96791bc4-2d83-4e8e-972f-657bdfb4174b service nova] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Updated VIF entry in instance network info cache for port 5661376f-2454-4693-a106-69cb78133962. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 767.665135] env[61867]: DEBUG nova.network.neutron [req-326dd822-5243-4c9a-86b6-ddd175306cb4 req-96791bc4-2d83-4e8e-972f-657bdfb4174b service nova] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Updating instance_info_cache with network_info: [{"id": "5661376f-2454-4693-a106-69cb78133962", "address": "fa:16:3e:7c:33:4d", "network": {"id": "998e68d3-8e78-4db8-8e53-859a2a9c494a", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1638393568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8451dfe23754279965eda6b2f2f37cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9297313e-7c50-4873-93d3-67284929163a", "external-id": "nsx-vlan-transportzone-620", "segmentation_id": 620, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5661376f-24", "ovs_interfaceid": "5661376f-2454-4693-a106-69cb78133962", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.667009] env[61867]: DEBUG oslo_concurrency.lockutils [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Acquiring lock "7c0badee-0e2b-4a13-89b1-e57df5c2c9ef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.738631] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "refresh_cache-81ada1e9-b100-44fd-8119-7ef8695e601f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.738768] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquired lock "refresh_cache-81ada1e9-b100-44fd-8119-7ef8695e601f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.738886] env[61867]: DEBUG nova.network.neutron [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 767.807939] env[61867]: DEBUG nova.network.neutron [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.853221] env[61867]: DEBUG oslo_concurrency.lockutils [None req-74a763b0-366a-4fb4-b4fa-53031a85b8ff tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Lock "7c0badee-0e2b-4a13-89b1-e57df5c2c9ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.488s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.854477] env[61867]: DEBUG oslo_concurrency.lockutils [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Lock "7c0badee-0e2b-4a13-89b1-e57df5c2c9ef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.187s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.854701] env[61867]: DEBUG oslo_concurrency.lockutils [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Acquiring lock "7c0badee-0e2b-4a13-89b1-e57df5c2c9ef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.854906] env[61867]: DEBUG oslo_concurrency.lockutils [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Lock "7c0badee-0e2b-4a13-89b1-e57df5c2c9ef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.855084] env[61867]: DEBUG oslo_concurrency.lockutils [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Lock "7c0badee-0e2b-4a13-89b1-e57df5c2c9ef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.857414] env[61867]: INFO nova.compute.manager [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Terminating instance [ 767.859160] env[61867]: DEBUG nova.compute.manager [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 767.859362] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 767.860337] env[61867]: DEBUG nova.network.neutron [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.862659] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d26524b5-deb5-47d4-a8a7-65ba5509f55b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.875657] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': task-1276409, 'name': Rename_Task, 'duration_secs': 0.142303} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.877641] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 767.877917] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 767.878133] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6dcf8000-bbe6-4f20-83e1-6c5980133008 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.879493] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ba560d2-6fee-4ce7-9da8-e85d7bf821b7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.885333] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Waiting for the task: (returnval){ [ 767.885333] env[61867]: value = "task-1276410" [ 767.885333] env[61867]: _type = "Task" [ 767.885333] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.892740] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525e54fb-03e4-86cd-a1af-b52de615baaa, 'name': SearchDatastore_Task, 'duration_secs': 0.010492} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.892740] env[61867]: DEBUG oslo_vmware.api [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Waiting for the task: (returnval){ [ 767.892740] env[61867]: value = "task-1276411" [ 767.892740] env[61867]: _type = "Task" [ 767.892740] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.893411] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.893685] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 767.893931] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.894092] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.894294] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 767.897151] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8d2c7233-8fb0-4f40-a8b9-d6dea927455d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.901733] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': task-1276410, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.906727] env[61867]: DEBUG oslo_vmware.api [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': task-1276411, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.908255] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 767.908430] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 767.909435] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c9bc8d4-a08c-4507-b33c-183ab972a70b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.914228] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 767.914228] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523bcf1b-3c04-e6f3-e56e-c63e70e4ab82" [ 767.914228] env[61867]: _type = "Task" [ 767.914228] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.923290] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523bcf1b-3c04-e6f3-e56e-c63e70e4ab82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.168365] env[61867]: DEBUG oslo_concurrency.lockutils [req-326dd822-5243-4c9a-86b6-ddd175306cb4 req-96791bc4-2d83-4e8e-972f-657bdfb4174b service nova] Releasing lock "refresh_cache-b5a6ff7c-2b48-4f82-ba53-ec1977736c8f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.274045] env[61867]: DEBUG nova.network.neutron [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.363166] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Releasing lock "refresh_cache-5e97f142-b5c0-4318-9a35-085234d1cb64" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.363489] env[61867]: DEBUG nova.compute.manager [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 768.363703] env[61867]: DEBUG nova.compute.manager [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 768.363887] env[61867]: DEBUG nova.network.neutron [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 768.365691] env[61867]: DEBUG nova.compute.manager [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 768.384596] env[61867]: DEBUG nova.network.neutron [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.394460] env[61867]: DEBUG nova.scheduler.client.report [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Refreshing inventories for resource provider 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 768.404096] env[61867]: DEBUG oslo_vmware.api [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': task-1276410, 'name': PowerOnVM_Task, 'duration_secs': 0.486639} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.404986] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 768.405212] env[61867]: INFO nova.compute.manager [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Took 6.34 seconds to spawn the instance on the hypervisor. [ 768.405388] env[61867]: DEBUG nova.compute.manager [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 768.406767] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac690c2-44dc-419b-abed-bd5620d71435 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.412183] env[61867]: DEBUG oslo_vmware.api [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': task-1276411, 'name': PowerOffVM_Task, 'duration_secs': 0.193175} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.413104] env[61867]: DEBUG nova.scheduler.client.report [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Updating ProviderTree inventory for provider 25720271-a549-4916-abe3-e5ed9b765889 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 768.413310] env[61867]: DEBUG nova.compute.provider_tree [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 768.415389] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 768.415564] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 768.416289] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-701b08cb-5e26-4eee-9c3e-493374f109ec {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.425166] env[61867]: DEBUG nova.scheduler.client.report [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Refreshing aggregate associations for resource provider 25720271-a549-4916-abe3-e5ed9b765889, aggregates: None {{(pid=61867) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 768.433404] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523bcf1b-3c04-e6f3-e56e-c63e70e4ab82, 'name': SearchDatastore_Task, 'duration_secs': 0.009282} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.434241] env[61867]: DEBUG nova.network.neutron [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Updating instance_info_cache with network_info: [{"id": "471a814b-d572-4c6b-bdbf-9e02e2091d77", "address": "fa:16:3e:60:d1:7d", "network": {"id": "998e68d3-8e78-4db8-8e53-859a2a9c494a", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1638393568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8451dfe23754279965eda6b2f2f37cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9297313e-7c50-4873-93d3-67284929163a", "external-id": "nsx-vlan-transportzone-620", "segmentation_id": 620, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap471a814b-d5", "ovs_interfaceid": "471a814b-d572-4c6b-bdbf-9e02e2091d77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.436389] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c91ce045-84ea-47e1-9010-98946a3ae973 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.441155] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 768.441155] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52fbe837-00ed-5348-fccc-a0036cb85379" [ 768.441155] env[61867]: _type = "Task" [ 768.441155] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.445798] env[61867]: DEBUG nova.scheduler.client.report [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Refreshing trait associations for resource provider 25720271-a549-4916-abe3-e5ed9b765889, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61867) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 768.451721] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52fbe837-00ed-5348-fccc-a0036cb85379, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.485319] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 768.485541] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 768.485720] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Deleting the datastore file [datastore2] 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 768.485971] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-59497c73-b293-4156-8eb1-efb4fe1bec09 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.493349] env[61867]: DEBUG oslo_vmware.api [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Waiting for the task: (returnval){ [ 768.493349] env[61867]: value = "task-1276413" [ 768.493349] env[61867]: _type = "Task" [ 768.493349] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.503540] env[61867]: DEBUG oslo_vmware.api [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': task-1276413, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.736102] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a247dbd1-8ec9-496a-9af9-5e7c0fa26415 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.741483] env[61867]: DEBUG nova.compute.manager [req-7aa5fd03-c5ff-4afb-8585-2ad0d45d40ed req-325ebcf9-83ad-4287-abf1-e36585d4dc71 service nova] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Received event network-vif-plugged-471a814b-d572-4c6b-bdbf-9e02e2091d77 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.741636] env[61867]: DEBUG oslo_concurrency.lockutils [req-7aa5fd03-c5ff-4afb-8585-2ad0d45d40ed req-325ebcf9-83ad-4287-abf1-e36585d4dc71 service nova] Acquiring lock "81ada1e9-b100-44fd-8119-7ef8695e601f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.741845] env[61867]: DEBUG oslo_concurrency.lockutils [req-7aa5fd03-c5ff-4afb-8585-2ad0d45d40ed req-325ebcf9-83ad-4287-abf1-e36585d4dc71 service nova] Lock "81ada1e9-b100-44fd-8119-7ef8695e601f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.742021] env[61867]: DEBUG oslo_concurrency.lockutils [req-7aa5fd03-c5ff-4afb-8585-2ad0d45d40ed req-325ebcf9-83ad-4287-abf1-e36585d4dc71 service nova] Lock "81ada1e9-b100-44fd-8119-7ef8695e601f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.742318] env[61867]: DEBUG nova.compute.manager [req-7aa5fd03-c5ff-4afb-8585-2ad0d45d40ed req-325ebcf9-83ad-4287-abf1-e36585d4dc71 service nova] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] No waiting events found dispatching network-vif-plugged-471a814b-d572-4c6b-bdbf-9e02e2091d77 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 768.742532] env[61867]: WARNING nova.compute.manager [req-7aa5fd03-c5ff-4afb-8585-2ad0d45d40ed req-325ebcf9-83ad-4287-abf1-e36585d4dc71 service nova] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Received unexpected event network-vif-plugged-471a814b-d572-4c6b-bdbf-9e02e2091d77 for instance with vm_state building and task_state spawning. [ 768.742695] env[61867]: DEBUG nova.compute.manager [req-7aa5fd03-c5ff-4afb-8585-2ad0d45d40ed req-325ebcf9-83ad-4287-abf1-e36585d4dc71 service nova] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Received event network-changed-471a814b-d572-4c6b-bdbf-9e02e2091d77 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.742854] env[61867]: DEBUG nova.compute.manager [req-7aa5fd03-c5ff-4afb-8585-2ad0d45d40ed req-325ebcf9-83ad-4287-abf1-e36585d4dc71 service nova] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Refreshing instance network info cache due to event network-changed-471a814b-d572-4c6b-bdbf-9e02e2091d77. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 768.743065] env[61867]: DEBUG oslo_concurrency.lockutils [req-7aa5fd03-c5ff-4afb-8585-2ad0d45d40ed req-325ebcf9-83ad-4287-abf1-e36585d4dc71 service nova] Acquiring lock "refresh_cache-81ada1e9-b100-44fd-8119-7ef8695e601f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.747869] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d12105-a4b3-4a5e-b9b5-5c6158b9bdcd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.778948] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac625a84-a8a8-4585-9969-50ab3c68a0d6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.786071] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e0def6-841e-4d30-98d4-ccd3fb95b5c6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.799965] env[61867]: DEBUG nova.compute.provider_tree [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.885430] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.887949] env[61867]: DEBUG nova.network.neutron [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.935959] env[61867]: INFO nova.compute.manager [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Took 35.15 seconds to build instance. [ 768.937371] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Releasing lock "refresh_cache-81ada1e9-b100-44fd-8119-7ef8695e601f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.937443] env[61867]: DEBUG nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Instance network_info: |[{"id": "471a814b-d572-4c6b-bdbf-9e02e2091d77", "address": "fa:16:3e:60:d1:7d", "network": {"id": "998e68d3-8e78-4db8-8e53-859a2a9c494a", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1638393568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8451dfe23754279965eda6b2f2f37cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9297313e-7c50-4873-93d3-67284929163a", "external-id": "nsx-vlan-transportzone-620", "segmentation_id": 620, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap471a814b-d5", "ovs_interfaceid": "471a814b-d572-4c6b-bdbf-9e02e2091d77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 768.937858] env[61867]: DEBUG oslo_concurrency.lockutils [req-7aa5fd03-c5ff-4afb-8585-2ad0d45d40ed req-325ebcf9-83ad-4287-abf1-e36585d4dc71 service nova] Acquired lock "refresh_cache-81ada1e9-b100-44fd-8119-7ef8695e601f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.938087] env[61867]: DEBUG nova.network.neutron [req-7aa5fd03-c5ff-4afb-8585-2ad0d45d40ed req-325ebcf9-83ad-4287-abf1-e36585d4dc71 service nova] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Refreshing network info cache for port 471a814b-d572-4c6b-bdbf-9e02e2091d77 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 768.939922] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:d1:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9297313e-7c50-4873-93d3-67284929163a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '471a814b-d572-4c6b-bdbf-9e02e2091d77', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 768.948261] env[61867]: DEBUG oslo.service.loopingcall [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.949413] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 768.952839] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1841c2f0-112f-4262-8d6a-5367b5b05871 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.972307] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52fbe837-00ed-5348-fccc-a0036cb85379, 'name': SearchDatastore_Task, 'duration_secs': 0.011401} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.973442] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.973690] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] b5a6ff7c-2b48-4f82-ba53-ec1977736c8f/b5a6ff7c-2b48-4f82-ba53-ec1977736c8f.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 768.973910] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 768.973910] env[61867]: value = "task-1276414" [ 768.973910] env[61867]: _type = "Task" [ 768.973910] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.974094] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-34b8cb5f-8875-4b82-bcfe-6cc6ab8ee71e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.983197] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276414, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.984561] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 768.984561] env[61867]: value = "task-1276415" [ 768.984561] env[61867]: _type = "Task" [ 768.984561] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.992274] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276415, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.001997] env[61867]: DEBUG oslo_vmware.api [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Task: {'id': task-1276413, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137714} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.002266] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 769.002434] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 769.002606] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 769.002846] env[61867]: INFO nova.compute.manager [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Took 1.14 seconds to destroy the instance on the hypervisor. [ 769.003066] env[61867]: DEBUG oslo.service.loopingcall [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 769.003262] env[61867]: DEBUG nova.compute.manager [-] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 769.003363] env[61867]: DEBUG nova.network.neutron [-] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 769.302897] env[61867]: DEBUG nova.scheduler.client.report [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 769.391924] env[61867]: INFO nova.compute.manager [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] [instance: 5e97f142-b5c0-4318-9a35-085234d1cb64] Took 1.03 seconds to deallocate network for instance. [ 769.439406] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1406885d-1d51-494b-af90-f32902941a05 tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Lock "d0db58a5-d773-4e45-a238-07cfc156bdd2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.351s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.493583] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276414, 'name': CreateVM_Task, 'duration_secs': 0.429751} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.494033] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 769.494836] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.495672] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.495672] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 769.499198] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-727f8c4f-3075-4312-b9af-36ecb75456af {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.500442] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276415, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.503683] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 769.503683] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52786671-ae93-9a4e-27df-2abcb85d2a9c" [ 769.503683] env[61867]: _type = "Task" [ 769.503683] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.511736] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52786671-ae93-9a4e-27df-2abcb85d2a9c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.645855] env[61867]: DEBUG nova.network.neutron [req-7aa5fd03-c5ff-4afb-8585-2ad0d45d40ed req-325ebcf9-83ad-4287-abf1-e36585d4dc71 service nova] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Updated VIF entry in instance network info cache for port 471a814b-d572-4c6b-bdbf-9e02e2091d77. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 769.646374] env[61867]: DEBUG nova.network.neutron [req-7aa5fd03-c5ff-4afb-8585-2ad0d45d40ed req-325ebcf9-83ad-4287-abf1-e36585d4dc71 service nova] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Updating instance_info_cache with network_info: [{"id": "471a814b-d572-4c6b-bdbf-9e02e2091d77", "address": "fa:16:3e:60:d1:7d", "network": {"id": "998e68d3-8e78-4db8-8e53-859a2a9c494a", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1638393568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8451dfe23754279965eda6b2f2f37cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9297313e-7c50-4873-93d3-67284929163a", "external-id": "nsx-vlan-transportzone-620", "segmentation_id": 620, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap471a814b-d5", "ovs_interfaceid": "471a814b-d572-4c6b-bdbf-9e02e2091d77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.778781] env[61867]: DEBUG nova.network.neutron [-] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.808544] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.809111] env[61867]: DEBUG nova.compute.manager [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 769.811853] env[61867]: DEBUG oslo_concurrency.lockutils [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.259s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.941689] env[61867]: DEBUG nova.compute.manager [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.995789] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276415, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.565729} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.996121] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] b5a6ff7c-2b48-4f82-ba53-ec1977736c8f/b5a6ff7c-2b48-4f82-ba53-ec1977736c8f.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 769.996345] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 769.996583] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-baa3608a-d34d-4401-b96d-ad05b39b191a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.004532] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 770.004532] env[61867]: value = "task-1276416" [ 770.004532] env[61867]: _type = "Task" [ 770.004532] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.016313] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52786671-ae93-9a4e-27df-2abcb85d2a9c, 'name': SearchDatastore_Task, 'duration_secs': 0.02058} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.019326] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.019599] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 770.019908] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.020141] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.021090] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 770.021090] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276416, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.021090] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3001327-5822-47b2-aa3e-ea7b18718ed2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.027697] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 770.027891] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 770.029470] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d335860e-c182-4fab-bf75-97f2f486778d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.037031] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 770.037031] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529c3f1b-c13a-e03f-616c-802fb1264c3a" [ 770.037031] env[61867]: _type = "Task" [ 770.037031] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.045470] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529c3f1b-c13a-e03f-616c-802fb1264c3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.149800] env[61867]: DEBUG oslo_concurrency.lockutils [req-7aa5fd03-c5ff-4afb-8585-2ad0d45d40ed req-325ebcf9-83ad-4287-abf1-e36585d4dc71 service nova] Releasing lock "refresh_cache-81ada1e9-b100-44fd-8119-7ef8695e601f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.281428] env[61867]: INFO nova.compute.manager [-] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Took 1.28 seconds to deallocate network for instance. [ 770.316048] env[61867]: DEBUG nova.compute.utils [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 770.320643] env[61867]: DEBUG nova.compute.manager [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 770.320826] env[61867]: DEBUG nova.network.neutron [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 770.366687] env[61867]: DEBUG nova.compute.manager [None req-ae904cc8-f4c8-4e2b-bec6-18b4249ddcef tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 770.367998] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-489c2ac2-32ae-4071-b716-06ee4edd943e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.379163] env[61867]: DEBUG nova.policy [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e9f1b09f23444730b91e1242f1e83652', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd700d9c347cf4401931229f6871e5c84', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 770.421989] env[61867]: INFO nova.scheduler.client.report [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Deleted allocations for instance 5e97f142-b5c0-4318-9a35-085234d1cb64 [ 770.461558] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.519358] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276416, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066507} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.519525] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 770.520329] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b35e251e-5d5e-4a96-aa56-883504138aa6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.536556] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Acquiring lock "d0db58a5-d773-4e45-a238-07cfc156bdd2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.536793] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Lock "d0db58a5-d773-4e45-a238-07cfc156bdd2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.536989] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Acquiring lock "d0db58a5-d773-4e45-a238-07cfc156bdd2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.537181] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Lock "d0db58a5-d773-4e45-a238-07cfc156bdd2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.537346] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Lock "d0db58a5-d773-4e45-a238-07cfc156bdd2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.547289] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] b5a6ff7c-2b48-4f82-ba53-ec1977736c8f/b5a6ff7c-2b48-4f82-ba53-ec1977736c8f.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 770.549969] env[61867]: INFO nova.compute.manager [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Terminating instance [ 770.551373] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9b51d72-6407-4030-98cb-1edb4abfccef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.565964] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Acquiring lock "refresh_cache-d0db58a5-d773-4e45-a238-07cfc156bdd2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.566135] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Acquired lock "refresh_cache-d0db58a5-d773-4e45-a238-07cfc156bdd2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.566303] env[61867]: DEBUG nova.network.neutron [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 770.576409] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529c3f1b-c13a-e03f-616c-802fb1264c3a, 'name': SearchDatastore_Task, 'duration_secs': 0.009458} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.580793] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 770.580793] env[61867]: value = "task-1276417" [ 770.580793] env[61867]: _type = "Task" [ 770.580793] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.581185] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ea7cfd5-70b7-4e87-8dec-b39d298e75df {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.591966] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276417, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.595725] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 770.595725] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52017479-61c8-c07a-fe6c-46956e85ccd4" [ 770.595725] env[61867]: _type = "Task" [ 770.595725] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.604399] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52017479-61c8-c07a-fe6c-46956e85ccd4, 'name': SearchDatastore_Task, 'duration_secs': 0.010762} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.606755] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.607017] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 81ada1e9-b100-44fd-8119-7ef8695e601f/81ada1e9-b100-44fd-8119-7ef8695e601f.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 770.607433] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7c0e9060-338d-4e2f-85be-04e366d6fcc0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.612973] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 770.612973] env[61867]: value = "task-1276418" [ 770.612973] env[61867]: _type = "Task" [ 770.612973] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.621430] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276418, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.664201] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1fb620a-e2f5-4074-b2d1-0fb188d7feae {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.671034] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c861be45-828f-4f55-bbd0-bf789dfe4f06 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.701692] env[61867]: DEBUG nova.network.neutron [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Successfully created port: d14dbc28-af77-4bdc-88e1-c56a5c320657 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 770.704052] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e31f95-017b-4d3e-b7df-91a8136f9c79 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.711368] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb8ccbf-9137-4090-83a6-035d4fad6cf6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.725590] env[61867]: DEBUG nova.compute.provider_tree [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.768774] env[61867]: DEBUG nova.compute.manager [req-99fba935-23b1-465b-9138-a6105d7d5fdc req-46ad22a7-17ae-4afe-a81f-55df0fb7e0f6 service nova] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Received event network-vif-deleted-bb4a3c79-631f-4359-927a-a2a7b2624aa9 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 770.788498] env[61867]: DEBUG oslo_concurrency.lockutils [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.824245] env[61867]: DEBUG nova.compute.manager [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 770.879366] env[61867]: INFO nova.compute.manager [None req-ae904cc8-f4c8-4e2b-bec6-18b4249ddcef tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] instance snapshotting [ 770.880079] env[61867]: DEBUG nova.objects.instance [None req-ae904cc8-f4c8-4e2b-bec6-18b4249ddcef tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Lazy-loading 'flavor' on Instance uuid d0db58a5-d773-4e45-a238-07cfc156bdd2 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 770.933048] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3569b7e2-38b4-49c6-9026-7f5cc9c0f8dc tempest-ServerDiagnosticsNegativeTest-943679031 tempest-ServerDiagnosticsNegativeTest-943679031-project-member] Lock "5e97f142-b5c0-4318-9a35-085234d1cb64" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 170.133s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.088523] env[61867]: DEBUG nova.network.neutron [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.098099] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276417, 'name': ReconfigVM_Task, 'duration_secs': 0.439497} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.098447] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Reconfigured VM instance instance-00000036 to attach disk [datastore2] b5a6ff7c-2b48-4f82-ba53-ec1977736c8f/b5a6ff7c-2b48-4f82-ba53-ec1977736c8f.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 771.099481] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5ceba488-81b9-4d3e-baed-813592980ea4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.106842] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 771.106842] env[61867]: value = "task-1276419" [ 771.106842] env[61867]: _type = "Task" [ 771.106842] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.117972] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276419, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.129467] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276418, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.152716] env[61867]: DEBUG nova.network.neutron [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.228406] env[61867]: DEBUG nova.scheduler.client.report [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.388397] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815ba577-f3ab-4e0d-84ba-57842934008f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.406832] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e416f6a5-3b65-47f8-a683-6a97becafcff {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.437339] env[61867]: DEBUG nova.compute.manager [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 771.616592] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276419, 'name': Rename_Task, 'duration_secs': 0.164614} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.616866] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 771.617172] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a98b5996-633f-460c-8728-35d19c9e1be2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.626765] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276418, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.844983} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.627821] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 81ada1e9-b100-44fd-8119-7ef8695e601f/81ada1e9-b100-44fd-8119-7ef8695e601f.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 771.628056] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 771.628336] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 771.628336] env[61867]: value = "task-1276420" [ 771.628336] env[61867]: _type = "Task" [ 771.628336] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.628507] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0369ad88-286d-420f-ac40-f79aeb5b7035 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.636857] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276420, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.637923] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 771.637923] env[61867]: value = "task-1276421" [ 771.637923] env[61867]: _type = "Task" [ 771.637923] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.644379] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276421, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.654966] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Releasing lock "refresh_cache-d0db58a5-d773-4e45-a238-07cfc156bdd2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.655386] env[61867]: DEBUG nova.compute.manager [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 771.655575] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 771.656329] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b4aca1d-b0ab-41ed-9d7a-2f02f7db8913 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.663117] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 771.663353] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8b4a6028-8d47-4211-b845-7806b866cb6b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.668744] env[61867]: DEBUG oslo_vmware.api [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Waiting for the task: (returnval){ [ 771.668744] env[61867]: value = "task-1276422" [ 771.668744] env[61867]: _type = "Task" [ 771.668744] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.675942] env[61867]: DEBUG oslo_vmware.api [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': task-1276422, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.734712] env[61867]: DEBUG oslo_concurrency.lockutils [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.923s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.735495] env[61867]: ERROR nova.compute.manager [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4c609ff8-9b47-4769-ba7e-0618fe97fd29, please check neutron logs for more information. [ 771.735495] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Traceback (most recent call last): [ 771.735495] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 771.735495] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] self.driver.spawn(context, instance, image_meta, [ 771.735495] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 771.735495] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 771.735495] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 771.735495] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] vm_ref = self.build_virtual_machine(instance, [ 771.735495] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 771.735495] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] vif_infos = vmwarevif.get_vif_info(self._session, [ 771.735495] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 771.735985] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] for vif in network_info: [ 771.735985] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 771.735985] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] return self._sync_wrapper(fn, *args, **kwargs) [ 771.735985] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 771.735985] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] self.wait() [ 771.735985] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 771.735985] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] self[:] = self._gt.wait() [ 771.735985] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 771.735985] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] return self._exit_event.wait() [ 771.735985] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 771.735985] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] current.throw(*self._exc) [ 771.735985] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 771.735985] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] result = function(*args, **kwargs) [ 771.736433] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 771.736433] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] return func(*args, **kwargs) [ 771.736433] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 771.736433] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] raise e [ 771.736433] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 771.736433] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] nwinfo = self.network_api.allocate_for_instance( [ 771.736433] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 771.736433] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] created_port_ids = self._update_ports_for_instance( [ 771.736433] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 771.736433] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] with excutils.save_and_reraise_exception(): [ 771.736433] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 771.736433] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] self.force_reraise() [ 771.736433] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 771.736894] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] raise self.value [ 771.736894] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 771.736894] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] updated_port = self._update_port( [ 771.736894] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 771.736894] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] _ensure_no_port_binding_failure(port) [ 771.736894] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 771.736894] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] raise exception.PortBindingFailed(port_id=port['id']) [ 771.736894] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] nova.exception.PortBindingFailed: Binding failed for port 4c609ff8-9b47-4769-ba7e-0618fe97fd29, please check neutron logs for more information. [ 771.736894] env[61867]: ERROR nova.compute.manager [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] [ 771.736894] env[61867]: DEBUG nova.compute.utils [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Binding failed for port 4c609ff8-9b47-4769-ba7e-0618fe97fd29, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 771.737863] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.347s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.739555] env[61867]: INFO nova.compute.claims [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.742400] env[61867]: DEBUG nova.compute.manager [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Build of instance 796ab4d3-ea9a-495d-9807-7780b1a7b0d6 was re-scheduled: Binding failed for port 4c609ff8-9b47-4769-ba7e-0618fe97fd29, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 771.742817] env[61867]: DEBUG nova.compute.manager [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 771.743085] env[61867]: DEBUG oslo_concurrency.lockutils [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Acquiring lock "refresh_cache-796ab4d3-ea9a-495d-9807-7780b1a7b0d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.743210] env[61867]: DEBUG oslo_concurrency.lockutils [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Acquired lock "refresh_cache-796ab4d3-ea9a-495d-9807-7780b1a7b0d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.743367] env[61867]: DEBUG nova.network.neutron [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 771.834398] env[61867]: DEBUG nova.compute.manager [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 771.858711] env[61867]: DEBUG nova.virt.hardware [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 771.858974] env[61867]: DEBUG nova.virt.hardware [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 771.859144] env[61867]: DEBUG nova.virt.hardware [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 771.859342] env[61867]: DEBUG nova.virt.hardware [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 771.859463] env[61867]: DEBUG nova.virt.hardware [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 771.859606] env[61867]: DEBUG nova.virt.hardware [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 771.859826] env[61867]: DEBUG nova.virt.hardware [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 771.859998] env[61867]: DEBUG nova.virt.hardware [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 771.860232] env[61867]: DEBUG nova.virt.hardware [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 771.860359] env[61867]: DEBUG nova.virt.hardware [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 771.860674] env[61867]: DEBUG nova.virt.hardware [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 771.861450] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d4ec7a3-2f3f-468c-bb00-ec0270caf678 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.870152] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7abbe588-e532-4451-a749-99ed1ed24f96 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.921449] env[61867]: DEBUG nova.compute.manager [None req-ae904cc8-f4c8-4e2b-bec6-18b4249ddcef tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Instance disappeared during snapshot {{(pid=61867) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 771.964030] env[61867]: DEBUG oslo_concurrency.lockutils [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.085649] env[61867]: DEBUG nova.compute.manager [None req-ae904cc8-f4c8-4e2b-bec6-18b4249ddcef tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Found 0 images (rotation: 2) {{(pid=61867) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 772.142366] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276420, 'name': PowerOnVM_Task} progress is 37%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.150767] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276421, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069029} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.151071] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 772.151844] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ec9724-3132-422e-9354-0f31d9c074c8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.174196] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 81ada1e9-b100-44fd-8119-7ef8695e601f/81ada1e9-b100-44fd-8119-7ef8695e601f.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 772.174196] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e85cdeff-a73c-4cb5-94ff-18d117817809 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.197980] env[61867]: DEBUG oslo_vmware.api [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': task-1276422, 'name': PowerOffVM_Task, 'duration_secs': 0.121482} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.199293] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 772.199474] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 772.199805] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 772.199805] env[61867]: value = "task-1276423" [ 772.199805] env[61867]: _type = "Task" [ 772.199805] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.200009] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fe9f6aae-19ec-416b-92d4-e0f55951d9fe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.209409] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276423, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.226442] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 772.226678] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 772.226960] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Deleting the datastore file [datastore2] d0db58a5-d773-4e45-a238-07cfc156bdd2 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 772.227286] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e26eea3-6b73-4c0d-9833-795c96efcf82 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.235041] env[61867]: DEBUG oslo_vmware.api [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Waiting for the task: (returnval){ [ 772.235041] env[61867]: value = "task-1276425" [ 772.235041] env[61867]: _type = "Task" [ 772.235041] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.243981] env[61867]: DEBUG oslo_vmware.api [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': task-1276425, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.264563] env[61867]: DEBUG nova.network.neutron [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.314674] env[61867]: DEBUG nova.compute.manager [req-079db736-da68-466c-82f4-4c19d66fcdfa req-8f715683-b8e5-4a45-80db-84903dd6f9cc service nova] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Received event network-vif-plugged-d14dbc28-af77-4bdc-88e1-c56a5c320657 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 772.315026] env[61867]: DEBUG oslo_concurrency.lockutils [req-079db736-da68-466c-82f4-4c19d66fcdfa req-8f715683-b8e5-4a45-80db-84903dd6f9cc service nova] Acquiring lock "c7203cea-2c2d-4a97-8ff4-3d33cd8443f7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.315410] env[61867]: DEBUG oslo_concurrency.lockutils [req-079db736-da68-466c-82f4-4c19d66fcdfa req-8f715683-b8e5-4a45-80db-84903dd6f9cc service nova] Lock "c7203cea-2c2d-4a97-8ff4-3d33cd8443f7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.315410] env[61867]: DEBUG oslo_concurrency.lockutils [req-079db736-da68-466c-82f4-4c19d66fcdfa req-8f715683-b8e5-4a45-80db-84903dd6f9cc service nova] Lock "c7203cea-2c2d-4a97-8ff4-3d33cd8443f7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.315513] env[61867]: DEBUG nova.compute.manager [req-079db736-da68-466c-82f4-4c19d66fcdfa req-8f715683-b8e5-4a45-80db-84903dd6f9cc service nova] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] No waiting events found dispatching network-vif-plugged-d14dbc28-af77-4bdc-88e1-c56a5c320657 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 772.316355] env[61867]: WARNING nova.compute.manager [req-079db736-da68-466c-82f4-4c19d66fcdfa req-8f715683-b8e5-4a45-80db-84903dd6f9cc service nova] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Received unexpected event network-vif-plugged-d14dbc28-af77-4bdc-88e1-c56a5c320657 for instance with vm_state building and task_state spawning. [ 772.318257] env[61867]: DEBUG nova.network.neutron [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.642304] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276420, 'name': PowerOnVM_Task, 'duration_secs': 0.969804} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.642623] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 772.642786] env[61867]: INFO nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Took 8.22 seconds to spawn the instance on the hypervisor. [ 772.642963] env[61867]: DEBUG nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 772.643743] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7858e9ba-addd-4459-9f67-91992242c2e9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.710468] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276423, 'name': ReconfigVM_Task, 'duration_secs': 0.295486} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.710839] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 81ada1e9-b100-44fd-8119-7ef8695e601f/81ada1e9-b100-44fd-8119-7ef8695e601f.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 772.711483] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b7bc4367-ce2f-45e4-854d-e3ea26839b30 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.717237] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 772.717237] env[61867]: value = "task-1276426" [ 772.717237] env[61867]: _type = "Task" [ 772.717237] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.726335] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276426, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.742870] env[61867]: DEBUG oslo_vmware.api [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Task: {'id': task-1276425, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.099782} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.743135] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 772.743327] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 772.743498] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 772.743681] env[61867]: INFO nova.compute.manager [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Took 1.09 seconds to destroy the instance on the hypervisor. [ 772.743936] env[61867]: DEBUG oslo.service.loopingcall [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.744147] env[61867]: DEBUG nova.compute.manager [-] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 772.744915] env[61867]: DEBUG nova.network.neutron [-] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 772.771634] env[61867]: DEBUG nova.network.neutron [-] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.797016] env[61867]: DEBUG nova.network.neutron [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Successfully updated port: d14dbc28-af77-4bdc-88e1-c56a5c320657 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 772.822492] env[61867]: DEBUG oslo_concurrency.lockutils [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Releasing lock "refresh_cache-796ab4d3-ea9a-495d-9807-7780b1a7b0d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.822851] env[61867]: DEBUG nova.compute.manager [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 772.823041] env[61867]: DEBUG nova.compute.manager [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 772.823212] env[61867]: DEBUG nova.network.neutron [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 772.827148] env[61867]: DEBUG nova.compute.manager [req-2bd5d645-3f1f-406f-9bcf-a2cfb3951645 req-437f3361-4214-4f4b-a4ea-bcc1a2846273 service nova] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Received event network-changed-d14dbc28-af77-4bdc-88e1-c56a5c320657 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 772.827356] env[61867]: DEBUG nova.compute.manager [req-2bd5d645-3f1f-406f-9bcf-a2cfb3951645 req-437f3361-4214-4f4b-a4ea-bcc1a2846273 service nova] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Refreshing instance network info cache due to event network-changed-d14dbc28-af77-4bdc-88e1-c56a5c320657. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 772.827568] env[61867]: DEBUG oslo_concurrency.lockutils [req-2bd5d645-3f1f-406f-9bcf-a2cfb3951645 req-437f3361-4214-4f4b-a4ea-bcc1a2846273 service nova] Acquiring lock "refresh_cache-c7203cea-2c2d-4a97-8ff4-3d33cd8443f7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.827708] env[61867]: DEBUG oslo_concurrency.lockutils [req-2bd5d645-3f1f-406f-9bcf-a2cfb3951645 req-437f3361-4214-4f4b-a4ea-bcc1a2846273 service nova] Acquired lock "refresh_cache-c7203cea-2c2d-4a97-8ff4-3d33cd8443f7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.827892] env[61867]: DEBUG nova.network.neutron [req-2bd5d645-3f1f-406f-9bcf-a2cfb3951645 req-437f3361-4214-4f4b-a4ea-bcc1a2846273 service nova] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Refreshing network info cache for port d14dbc28-af77-4bdc-88e1-c56a5c320657 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 772.847233] env[61867]: DEBUG nova.network.neutron [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.076019] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10aef8bd-88ac-426b-a261-88b9fcf38aea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.082409] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24965e5-566f-4aac-b06b-c109a1b1e5dc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.114083] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac4447d-c991-4d33-bb67-37f6806f99bb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.122759] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e692914b-623c-439f-bbf1-6d91818e56f2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.136600] env[61867]: DEBUG nova.compute.provider_tree [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.161292] env[61867]: INFO nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Took 36.92 seconds to build instance. [ 773.226621] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276426, 'name': Rename_Task, 'duration_secs': 0.130053} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.227395] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 773.227693] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46c4c59d-02ec-46e1-b2b0-de3a715c4eba {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.234053] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 773.234053] env[61867]: value = "task-1276427" [ 773.234053] env[61867]: _type = "Task" [ 773.234053] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.241196] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276427, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.274248] env[61867]: DEBUG nova.network.neutron [-] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.299120] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Acquiring lock "refresh_cache-c7203cea-2c2d-4a97-8ff4-3d33cd8443f7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.349517] env[61867]: DEBUG nova.network.neutron [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.369106] env[61867]: DEBUG nova.network.neutron [req-2bd5d645-3f1f-406f-9bcf-a2cfb3951645 req-437f3361-4214-4f4b-a4ea-bcc1a2846273 service nova] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.458437] env[61867]: DEBUG nova.network.neutron [req-2bd5d645-3f1f-406f-9bcf-a2cfb3951645 req-437f3361-4214-4f4b-a4ea-bcc1a2846273 service nova] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.643274] env[61867]: DEBUG nova.scheduler.client.report [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.663082] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "b5a6ff7c-2b48-4f82-ba53-ec1977736c8f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.025s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.744569] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276427, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.779330] env[61867]: INFO nova.compute.manager [-] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Took 1.03 seconds to deallocate network for instance. [ 773.852161] env[61867]: INFO nova.compute.manager [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 796ab4d3-ea9a-495d-9807-7780b1a7b0d6] Took 1.03 seconds to deallocate network for instance. [ 773.961529] env[61867]: DEBUG oslo_concurrency.lockutils [req-2bd5d645-3f1f-406f-9bcf-a2cfb3951645 req-437f3361-4214-4f4b-a4ea-bcc1a2846273 service nova] Releasing lock "refresh_cache-c7203cea-2c2d-4a97-8ff4-3d33cd8443f7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.961932] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Acquired lock "refresh_cache-c7203cea-2c2d-4a97-8ff4-3d33cd8443f7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.962115] env[61867]: DEBUG nova.network.neutron [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 774.150091] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.150680] env[61867]: DEBUG nova.compute.manager [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 774.156337] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 28.826s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.156535] env[61867]: DEBUG nova.objects.instance [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61867) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 774.166147] env[61867]: DEBUG nova.compute.manager [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 774.250516] env[61867]: DEBUG oslo_vmware.api [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276427, 'name': PowerOnVM_Task, 'duration_secs': 0.921428} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.251458] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 774.251751] env[61867]: INFO nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Took 6.97 seconds to spawn the instance on the hypervisor. [ 774.252038] env[61867]: DEBUG nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 774.252965] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c857118-db56-4804-9423-2075078b0889 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.287052] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.509291] env[61867]: DEBUG nova.network.neutron [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.662334] env[61867]: DEBUG nova.network.neutron [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Updating instance_info_cache with network_info: [{"id": "d14dbc28-af77-4bdc-88e1-c56a5c320657", "address": "fa:16:3e:aa:b9:7e", "network": {"id": "2d4ab52c-144f-45a1-9d4f-afd85fc30404", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.247", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "84601e35c8e8487cb78fc16a2536a4c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d47d5e1d-e66d-4f2c-83e6-d5e78c2b767d", "external-id": "nsx-vlan-transportzone-109", "segmentation_id": 109, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd14dbc28-af", "ovs_interfaceid": "d14dbc28-af77-4bdc-88e1-c56a5c320657", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.668892] env[61867]: DEBUG nova.compute.utils [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 774.675516] env[61867]: DEBUG nova.compute.manager [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 774.675698] env[61867]: DEBUG nova.network.neutron [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 774.702601] env[61867]: DEBUG oslo_concurrency.lockutils [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.720138] env[61867]: DEBUG nova.policy [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6cf013ebfa3346d1a7b00dfc49254d12', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd1d8c9b79b3a44e6ac62126ab3ada96b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 774.771409] env[61867]: INFO nova.compute.manager [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Took 37.26 seconds to build instance. [ 774.889408] env[61867]: INFO nova.scheduler.client.report [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Deleted allocations for instance 796ab4d3-ea9a-495d-9807-7780b1a7b0d6 [ 775.021510] env[61867]: DEBUG nova.network.neutron [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Successfully created port: fdee8656-849f-4969-8197-bce85b819979 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 775.180020] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Releasing lock "refresh_cache-c7203cea-2c2d-4a97-8ff4-3d33cd8443f7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.180020] env[61867]: DEBUG nova.compute.manager [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Instance network_info: |[{"id": "d14dbc28-af77-4bdc-88e1-c56a5c320657", "address": "fa:16:3e:aa:b9:7e", "network": {"id": "2d4ab52c-144f-45a1-9d4f-afd85fc30404", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.247", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "84601e35c8e8487cb78fc16a2536a4c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d47d5e1d-e66d-4f2c-83e6-d5e78c2b767d", "external-id": "nsx-vlan-transportzone-109", "segmentation_id": 109, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd14dbc28-af", "ovs_interfaceid": "d14dbc28-af77-4bdc-88e1-c56a5c320657", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 775.180405] env[61867]: DEBUG nova.compute.manager [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 775.182044] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:aa:b9:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd47d5e1d-e66d-4f2c-83e6-d5e78c2b767d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd14dbc28-af77-4bdc-88e1-c56a5c320657', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 775.191583] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Creating folder: Project (d700d9c347cf4401931229f6871e5c84). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 775.192988] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0eaa9bac-76db-4e4a-899b-8a562723bd63 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.037s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.194693] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f3e481c8-3198-4a05-b99e-9138dfc17677 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.200467] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.173s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.210671] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Created folder: Project (d700d9c347cf4401931229f6871e5c84) in parent group-v274258. [ 775.211462] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Creating folder: Instances. Parent ref: group-v274294. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 775.211462] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-10c0de3d-fbda-4bd6-8c3a-a7b561f31df1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.223562] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Created folder: Instances in parent group-v274294. [ 775.223836] env[61867]: DEBUG oslo.service.loopingcall [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 775.224907] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 775.224907] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f7d83f81-5d75-4688-b08e-6aff6d4999bc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.257238] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 775.257238] env[61867]: value = "task-1276430" [ 775.257238] env[61867]: _type = "Task" [ 775.257238] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.267147] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276430, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.274942] env[61867]: DEBUG oslo_concurrency.lockutils [None req-30fe1c3c-ccbc-4144-9c72-267d12ef18e8 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "81ada1e9-b100-44fd-8119-7ef8695e601f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.601s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.400055] env[61867]: DEBUG oslo_concurrency.lockutils [None req-27bd2f38-cb84-4bd2-a836-776b3115371c tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Lock "796ab4d3-ea9a-495d-9807-7780b1a7b0d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 173.598s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.766954] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276430, 'name': CreateVM_Task, 'duration_secs': 0.331283} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.769684] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 775.770585] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.770853] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.771224] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 775.771358] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87aa135a-397f-46e6-a558-a05be7530fa8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.776521] env[61867]: DEBUG nova.compute.manager [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 775.779636] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Waiting for the task: (returnval){ [ 775.779636] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521fbb96-763d-d2e6-4a6c-a07af9568b39" [ 775.779636] env[61867]: _type = "Task" [ 775.779636] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.790990] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521fbb96-763d-d2e6-4a6c-a07af9568b39, 'name': SearchDatastore_Task, 'duration_secs': 0.012311} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.791629] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.791629] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 775.792107] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.792107] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.792252] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 775.792541] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff00b2d2-d355-4d54-83de-307d0b7af454 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.802571] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 775.802808] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 775.803605] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed50ee73-e96c-4eff-996f-d561455f85f7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.809593] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Waiting for the task: (returnval){ [ 775.809593] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520325b4-eb53-c44b-63b6-f85b07505683" [ 775.809593] env[61867]: _type = "Task" [ 775.809593] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.826937] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520325b4-eb53-c44b-63b6-f85b07505683, 'name': SearchDatastore_Task, 'duration_secs': 0.009553} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.829031] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a576f30d-957d-49be-836f-3c5a184df5cb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.835631] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Waiting for the task: (returnval){ [ 775.835631] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5293a4f8-0875-17d4-c281-620ac847f503" [ 775.835631] env[61867]: _type = "Task" [ 775.835631] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.844601] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5293a4f8-0875-17d4-c281-620ac847f503, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.902400] env[61867]: DEBUG nova.compute.manager [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 775.913318] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "9aab8852-addb-49e6-a59b-fa9bffc7733b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.913546] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "9aab8852-addb-49e6-a59b-fa9bffc7733b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.059645] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-509ec5e0-2ac7-4a59-8233-e98ff6306614 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.067581] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80833510-589a-47b7-a322-7762906fe27e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.098483] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dc219bd-b1de-4c9c-a130-c17660d4c78b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.105794] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e358e20d-cb31-47af-9fdf-3e86be41020e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.120678] env[61867]: DEBUG nova.compute.provider_tree [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.168243] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "b5a6ff7c-2b48-4f82-ba53-ec1977736c8f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.168523] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "b5a6ff7c-2b48-4f82-ba53-ec1977736c8f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.168739] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "b5a6ff7c-2b48-4f82-ba53-ec1977736c8f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.168923] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "b5a6ff7c-2b48-4f82-ba53-ec1977736c8f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.169107] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "b5a6ff7c-2b48-4f82-ba53-ec1977736c8f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.171536] env[61867]: INFO nova.compute.manager [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Terminating instance [ 776.173443] env[61867]: DEBUG nova.compute.manager [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 776.173630] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 776.174440] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d4849a-55de-4629-a4c5-bcfb3476458e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.182988] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 776.183267] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-484c42dc-4ff8-42ea-93ef-592bdb72b8ea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.189296] env[61867]: DEBUG oslo_vmware.api [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 776.189296] env[61867]: value = "task-1276431" [ 776.189296] env[61867]: _type = "Task" [ 776.189296] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.197220] env[61867]: DEBUG oslo_vmware.api [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276431, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.201493] env[61867]: DEBUG nova.compute.manager [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 776.228415] env[61867]: DEBUG nova.virt.hardware [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 776.228415] env[61867]: DEBUG nova.virt.hardware [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 776.228415] env[61867]: DEBUG nova.virt.hardware [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 776.228813] env[61867]: DEBUG nova.virt.hardware [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 776.228813] env[61867]: DEBUG nova.virt.hardware [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 776.228813] env[61867]: DEBUG nova.virt.hardware [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 776.228945] env[61867]: DEBUG nova.virt.hardware [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 776.229108] env[61867]: DEBUG nova.virt.hardware [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 776.229340] env[61867]: DEBUG nova.virt.hardware [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 776.229520] env[61867]: DEBUG nova.virt.hardware [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 776.229691] env[61867]: DEBUG nova.virt.hardware [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 776.230582] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994b0968-7a4a-4f6a-9bc5-213978acedef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.238608] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3e49fe-e8ee-43cf-9697-66cfd40ff455 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.299858] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.346183] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5293a4f8-0875-17d4-c281-620ac847f503, 'name': SearchDatastore_Task, 'duration_secs': 0.009571} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.346516] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.347130] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] c7203cea-2c2d-4a97-8ff4-3d33cd8443f7/c7203cea-2c2d-4a97-8ff4-3d33cd8443f7.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 776.347130] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c4d2c025-b7f1-4522-9d3a-6ad07fa620a3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.349374] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "81ada1e9-b100-44fd-8119-7ef8695e601f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.349631] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "81ada1e9-b100-44fd-8119-7ef8695e601f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.349941] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "81ada1e9-b100-44fd-8119-7ef8695e601f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.350243] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "81ada1e9-b100-44fd-8119-7ef8695e601f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.350761] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "81ada1e9-b100-44fd-8119-7ef8695e601f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.353014] env[61867]: INFO nova.compute.manager [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Terminating instance [ 776.355946] env[61867]: DEBUG nova.compute.manager [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 776.356209] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 776.357055] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c354f50-8c5e-4e73-9501-990cf10dee72 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.362140] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Waiting for the task: (returnval){ [ 776.362140] env[61867]: value = "task-1276432" [ 776.362140] env[61867]: _type = "Task" [ 776.362140] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.367999] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 776.368637] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d3f02b7c-5ebd-4075-8ec3-9731fb0ed6f0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.373596] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': task-1276432, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.378303] env[61867]: DEBUG oslo_vmware.api [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 776.378303] env[61867]: value = "task-1276433" [ 776.378303] env[61867]: _type = "Task" [ 776.378303] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.391986] env[61867]: DEBUG oslo_vmware.api [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276433, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.425549] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.519206] env[61867]: DEBUG nova.compute.manager [req-20310235-fe50-4e94-a981-6f921f0cd893 req-d47c4e50-5782-49f3-8ff3-899bd3a428f0 service nova] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Received event network-vif-plugged-fdee8656-849f-4969-8197-bce85b819979 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.519282] env[61867]: DEBUG oslo_concurrency.lockutils [req-20310235-fe50-4e94-a981-6f921f0cd893 req-d47c4e50-5782-49f3-8ff3-899bd3a428f0 service nova] Acquiring lock "f0b6273f-4177-44a3-8dd1-f65faeb07539-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.519495] env[61867]: DEBUG oslo_concurrency.lockutils [req-20310235-fe50-4e94-a981-6f921f0cd893 req-d47c4e50-5782-49f3-8ff3-899bd3a428f0 service nova] Lock "f0b6273f-4177-44a3-8dd1-f65faeb07539-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.519666] env[61867]: DEBUG oslo_concurrency.lockutils [req-20310235-fe50-4e94-a981-6f921f0cd893 req-d47c4e50-5782-49f3-8ff3-899bd3a428f0 service nova] Lock "f0b6273f-4177-44a3-8dd1-f65faeb07539-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.519892] env[61867]: DEBUG nova.compute.manager [req-20310235-fe50-4e94-a981-6f921f0cd893 req-d47c4e50-5782-49f3-8ff3-899bd3a428f0 service nova] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] No waiting events found dispatching network-vif-plugged-fdee8656-849f-4969-8197-bce85b819979 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 776.520245] env[61867]: WARNING nova.compute.manager [req-20310235-fe50-4e94-a981-6f921f0cd893 req-d47c4e50-5782-49f3-8ff3-899bd3a428f0 service nova] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Received unexpected event network-vif-plugged-fdee8656-849f-4969-8197-bce85b819979 for instance with vm_state building and task_state spawning. [ 776.621576] env[61867]: DEBUG nova.network.neutron [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Successfully updated port: fdee8656-849f-4969-8197-bce85b819979 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 776.623895] env[61867]: DEBUG nova.scheduler.client.report [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.700529] env[61867]: DEBUG oslo_vmware.api [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276431, 'name': PowerOffVM_Task, 'duration_secs': 0.212056} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.700890] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 776.701094] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 776.701363] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8b4f86a3-8679-46f6-915b-22028dd4b9ef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.759231] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 776.759542] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 776.759997] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Deleting the datastore file [datastore2] b5a6ff7c-2b48-4f82-ba53-ec1977736c8f {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 776.760242] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-950ccafb-7462-41e2-851b-6bd3f6a995c1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.766814] env[61867]: DEBUG oslo_vmware.api [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 776.766814] env[61867]: value = "task-1276435" [ 776.766814] env[61867]: _type = "Task" [ 776.766814] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.774956] env[61867]: DEBUG oslo_vmware.api [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276435, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.874183] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': task-1276432, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481119} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.874316] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] c7203cea-2c2d-4a97-8ff4-3d33cd8443f7/c7203cea-2c2d-4a97-8ff4-3d33cd8443f7.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 776.874522] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 776.874763] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-01a2ca9d-ea54-4ed1-b644-61d2dafb99a4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.880929] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Waiting for the task: (returnval){ [ 776.880929] env[61867]: value = "task-1276436" [ 776.880929] env[61867]: _type = "Task" [ 776.880929] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.891655] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': task-1276436, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.894589] env[61867]: DEBUG oslo_vmware.api [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276433, 'name': PowerOffVM_Task, 'duration_secs': 0.195411} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.895124] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 776.895124] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 776.895213] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2caf1089-34ed-4635-85e3-04706bf1ffa9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.952582] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 776.952582] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 776.952582] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Deleting the datastore file [datastore1] 81ada1e9-b100-44fd-8119-7ef8695e601f {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 776.952582] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-37ba62b5-ce01-4d5b-b314-f1b7fa30f6c5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.958690] env[61867]: DEBUG oslo_vmware.api [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for the task: (returnval){ [ 776.958690] env[61867]: value = "task-1276438" [ 776.958690] env[61867]: _type = "Task" [ 776.958690] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.965745] env[61867]: DEBUG oslo_vmware.api [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276438, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.133461] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Acquiring lock "refresh_cache-f0b6273f-4177-44a3-8dd1-f65faeb07539" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.133613] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Acquired lock "refresh_cache-f0b6273f-4177-44a3-8dd1-f65faeb07539" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.134149] env[61867]: DEBUG nova.network.neutron [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 777.135587] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.939s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.136251] env[61867]: ERROR nova.compute.manager [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 35552c32-4d31-4fa5-a23d-e2dc0035d263, please check neutron logs for more information. [ 777.136251] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Traceback (most recent call last): [ 777.136251] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 777.136251] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] self.driver.spawn(context, instance, image_meta, [ 777.136251] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 777.136251] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] self._vmops.spawn(context, instance, image_meta, injected_files, [ 777.136251] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 777.136251] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] vm_ref = self.build_virtual_machine(instance, [ 777.136251] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 777.136251] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] vif_infos = vmwarevif.get_vif_info(self._session, [ 777.136251] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 777.136500] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] for vif in network_info: [ 777.136500] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 777.136500] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] return self._sync_wrapper(fn, *args, **kwargs) [ 777.136500] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 777.136500] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] self.wait() [ 777.136500] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 777.136500] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] self[:] = self._gt.wait() [ 777.136500] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 777.136500] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] return self._exit_event.wait() [ 777.136500] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 777.136500] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] current.throw(*self._exc) [ 777.136500] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 777.136500] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] result = function(*args, **kwargs) [ 777.136763] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 777.136763] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] return func(*args, **kwargs) [ 777.136763] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 777.136763] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] raise e [ 777.136763] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.136763] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] nwinfo = self.network_api.allocate_for_instance( [ 777.136763] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 777.136763] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] created_port_ids = self._update_ports_for_instance( [ 777.136763] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 777.136763] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] with excutils.save_and_reraise_exception(): [ 777.136763] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.136763] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] self.force_reraise() [ 777.136763] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.137248] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] raise self.value [ 777.137248] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 777.137248] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] updated_port = self._update_port( [ 777.137248] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.137248] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] _ensure_no_port_binding_failure(port) [ 777.137248] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.137248] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] raise exception.PortBindingFailed(port_id=port['id']) [ 777.137248] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] nova.exception.PortBindingFailed: Binding failed for port 35552c32-4d31-4fa5-a23d-e2dc0035d263, please check neutron logs for more information. [ 777.137248] env[61867]: ERROR nova.compute.manager [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] [ 777.137248] env[61867]: DEBUG nova.compute.utils [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Binding failed for port 35552c32-4d31-4fa5-a23d-e2dc0035d263, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 777.138531] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.354s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.141715] env[61867]: DEBUG nova.compute.manager [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Build of instance 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce was re-scheduled: Binding failed for port 35552c32-4d31-4fa5-a23d-e2dc0035d263, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 777.142212] env[61867]: DEBUG nova.compute.manager [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 777.142451] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Acquiring lock "refresh_cache-8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.142633] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Acquired lock "refresh_cache-8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.142821] env[61867]: DEBUG nova.network.neutron [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 777.277890] env[61867]: DEBUG oslo_vmware.api [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276435, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.224579} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.277890] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 777.277890] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 777.277890] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 777.277890] env[61867]: INFO nova.compute.manager [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Took 1.10 seconds to destroy the instance on the hypervisor. [ 777.278431] env[61867]: DEBUG oslo.service.loopingcall [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 777.278431] env[61867]: DEBUG nova.compute.manager [-] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 777.278431] env[61867]: DEBUG nova.network.neutron [-] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 777.390900] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': task-1276436, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087447} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.391200] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 777.391990] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-859972d5-c8ce-437f-9511-ab8cbe3caee6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.413802] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] c7203cea-2c2d-4a97-8ff4-3d33cd8443f7/c7203cea-2c2d-4a97-8ff4-3d33cd8443f7.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 777.414088] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4297d731-743b-437b-8505-f408d963885a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.433613] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Waiting for the task: (returnval){ [ 777.433613] env[61867]: value = "task-1276439" [ 777.433613] env[61867]: _type = "Task" [ 777.433613] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.442366] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': task-1276439, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.466814] env[61867]: DEBUG oslo_vmware.api [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Task: {'id': task-1276438, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187489} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.467079] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 777.467263] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 777.467439] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 777.467611] env[61867]: INFO nova.compute.manager [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Took 1.11 seconds to destroy the instance on the hypervisor. [ 777.467857] env[61867]: DEBUG oslo.service.loopingcall [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 777.468049] env[61867]: DEBUG nova.compute.manager [-] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 777.468133] env[61867]: DEBUG nova.network.neutron [-] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 777.675183] env[61867]: DEBUG nova.network.neutron [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.689446] env[61867]: DEBUG nova.network.neutron [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.764858] env[61867]: DEBUG nova.network.neutron [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.944027] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': task-1276439, 'name': ReconfigVM_Task, 'duration_secs': 0.444746} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.946358] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Reconfigured VM instance instance-00000038 to attach disk [datastore2] c7203cea-2c2d-4a97-8ff4-3d33cd8443f7/c7203cea-2c2d-4a97-8ff4-3d33cd8443f7.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 777.947150] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b0cac572-f5ee-4f15-aeab-fad11160da70 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.953176] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Waiting for the task: (returnval){ [ 777.953176] env[61867]: value = "task-1276440" [ 777.953176] env[61867]: _type = "Task" [ 777.953176] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.963022] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': task-1276440, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.992622] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21744f68-2289-46e8-8e3d-22bd6f3c53cf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.999740] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6abdb56d-db89-40ba-9442-5dea02bf2ab2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.003570] env[61867]: DEBUG nova.network.neutron [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Updating instance_info_cache with network_info: [{"id": "fdee8656-849f-4969-8197-bce85b819979", "address": "fa:16:3e:15:81:ae", "network": {"id": "b284d3b5-fef2-4159-a0d1-d44cc9d08c5f", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1659399239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1d8c9b79b3a44e6ac62126ab3ada96b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdee8656-84", "ovs_interfaceid": "fdee8656-849f-4969-8197-bce85b819979", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.031768] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2134906-575c-4612-a7ff-a462389dec8b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.039976] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb9c7836-8136-4a5d-873a-eded18128c08 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.053021] env[61867]: DEBUG nova.network.neutron [-] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.054643] env[61867]: DEBUG nova.compute.provider_tree [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.234724] env[61867]: DEBUG nova.network.neutron [-] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.271043] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Releasing lock "refresh_cache-8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.271043] env[61867]: DEBUG nova.compute.manager [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 778.271217] env[61867]: DEBUG nova.compute.manager [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 778.271355] env[61867]: DEBUG nova.network.neutron [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 778.286896] env[61867]: DEBUG nova.network.neutron [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.463847] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': task-1276440, 'name': Rename_Task, 'duration_secs': 0.182856} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.463847] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 778.464083] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2535515a-0152-448f-ae28-2809eaeceedd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.471411] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Waiting for the task: (returnval){ [ 778.471411] env[61867]: value = "task-1276441" [ 778.471411] env[61867]: _type = "Task" [ 778.471411] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.479995] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': task-1276441, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.506635] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Releasing lock "refresh_cache-f0b6273f-4177-44a3-8dd1-f65faeb07539" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.506984] env[61867]: DEBUG nova.compute.manager [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Instance network_info: |[{"id": "fdee8656-849f-4969-8197-bce85b819979", "address": "fa:16:3e:15:81:ae", "network": {"id": "b284d3b5-fef2-4159-a0d1-d44cc9d08c5f", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1659399239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1d8c9b79b3a44e6ac62126ab3ada96b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdee8656-84", "ovs_interfaceid": "fdee8656-849f-4969-8197-bce85b819979", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 778.507424] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:81:ae', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5ba07329-1d3e-4ba8-8774-d029262318c4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fdee8656-849f-4969-8197-bce85b819979', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 778.515125] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Creating folder: Project (d1d8c9b79b3a44e6ac62126ab3ada96b). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 778.515393] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-04bd3692-29ce-4e62-b822-1d50779d81fe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.525108] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Created folder: Project (d1d8c9b79b3a44e6ac62126ab3ada96b) in parent group-v274258. [ 778.525294] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Creating folder: Instances. Parent ref: group-v274297. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 778.525515] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-253ff649-dc2c-40f4-8280-647fe4f4d81f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.533760] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Created folder: Instances in parent group-v274297. [ 778.533987] env[61867]: DEBUG oslo.service.loopingcall [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.534212] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 778.534528] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7373dd0b-89b9-4655-b381-5bab8a1fc0b1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.555175] env[61867]: DEBUG nova.compute.manager [req-a45a47f0-5bc1-4fa5-8a6c-0e9421713f87 req-23294e6d-4c5d-4771-ae73-2e2eb4780ec0 service nova] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Received event network-changed-fdee8656-849f-4969-8197-bce85b819979 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 778.555661] env[61867]: DEBUG nova.compute.manager [req-a45a47f0-5bc1-4fa5-8a6c-0e9421713f87 req-23294e6d-4c5d-4771-ae73-2e2eb4780ec0 service nova] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Refreshing instance network info cache due to event network-changed-fdee8656-849f-4969-8197-bce85b819979. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 778.555661] env[61867]: DEBUG oslo_concurrency.lockutils [req-a45a47f0-5bc1-4fa5-8a6c-0e9421713f87 req-23294e6d-4c5d-4771-ae73-2e2eb4780ec0 service nova] Acquiring lock "refresh_cache-f0b6273f-4177-44a3-8dd1-f65faeb07539" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.555749] env[61867]: DEBUG oslo_concurrency.lockutils [req-a45a47f0-5bc1-4fa5-8a6c-0e9421713f87 req-23294e6d-4c5d-4771-ae73-2e2eb4780ec0 service nova] Acquired lock "refresh_cache-f0b6273f-4177-44a3-8dd1-f65faeb07539" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.555863] env[61867]: DEBUG nova.network.neutron [req-a45a47f0-5bc1-4fa5-8a6c-0e9421713f87 req-23294e6d-4c5d-4771-ae73-2e2eb4780ec0 service nova] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Refreshing network info cache for port fdee8656-849f-4969-8197-bce85b819979 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 778.558389] env[61867]: INFO nova.compute.manager [-] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Took 1.28 seconds to deallocate network for instance. [ 778.559088] env[61867]: DEBUG nova.scheduler.client.report [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 778.561941] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 778.561941] env[61867]: value = "task-1276444" [ 778.561941] env[61867]: _type = "Task" [ 778.561941] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.576029] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276444, 'name': CreateVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.737638] env[61867]: INFO nova.compute.manager [-] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Took 1.27 seconds to deallocate network for instance. [ 778.789571] env[61867]: DEBUG nova.network.neutron [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.981407] env[61867]: DEBUG oslo_vmware.api [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': task-1276441, 'name': PowerOnVM_Task, 'duration_secs': 0.462314} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.981682] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 778.981908] env[61867]: INFO nova.compute.manager [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Took 7.15 seconds to spawn the instance on the hypervisor. [ 778.982103] env[61867]: DEBUG nova.compute.manager [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 778.982860] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-980bd760-8c53-46a7-ac1e-c943c35d3e44 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.065471] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.927s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.066136] env[61867]: ERROR nova.compute.manager [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3b1112dd-2d7d-453e-ad9e-677480ae22c4, please check neutron logs for more information. [ 779.066136] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Traceback (most recent call last): [ 779.066136] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 779.066136] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] self.driver.spawn(context, instance, image_meta, [ 779.066136] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 779.066136] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 779.066136] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 779.066136] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] vm_ref = self.build_virtual_machine(instance, [ 779.066136] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 779.066136] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] vif_infos = vmwarevif.get_vif_info(self._session, [ 779.066136] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 779.066470] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] for vif in network_info: [ 779.066470] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 779.066470] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] return self._sync_wrapper(fn, *args, **kwargs) [ 779.066470] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 779.066470] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] self.wait() [ 779.066470] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 779.066470] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] self[:] = self._gt.wait() [ 779.066470] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 779.066470] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] return self._exit_event.wait() [ 779.066470] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 779.066470] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] result = hub.switch() [ 779.066470] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 779.066470] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] return self.greenlet.switch() [ 779.066954] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.066954] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] result = function(*args, **kwargs) [ 779.066954] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 779.066954] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] return func(*args, **kwargs) [ 779.066954] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.066954] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] raise e [ 779.066954] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.066954] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] nwinfo = self.network_api.allocate_for_instance( [ 779.066954] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.066954] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] created_port_ids = self._update_ports_for_instance( [ 779.066954] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.066954] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] with excutils.save_and_reraise_exception(): [ 779.066954] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.067317] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] self.force_reraise() [ 779.067317] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.067317] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] raise self.value [ 779.067317] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.067317] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] updated_port = self._update_port( [ 779.067317] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.067317] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] _ensure_no_port_binding_failure(port) [ 779.067317] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.067317] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] raise exception.PortBindingFailed(port_id=port['id']) [ 779.067317] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] nova.exception.PortBindingFailed: Binding failed for port 3b1112dd-2d7d-453e-ad9e-677480ae22c4, please check neutron logs for more information. [ 779.067317] env[61867]: ERROR nova.compute.manager [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] [ 779.067677] env[61867]: DEBUG nova.compute.utils [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Binding failed for port 3b1112dd-2d7d-453e-ad9e-677480ae22c4, please check neutron logs for more information. {{(pid=61867) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 779.068010] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.258s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.068949] env[61867]: DEBUG nova.objects.instance [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lazy-loading 'resources' on Instance uuid 8656c3b8-b7b1-442a-9719-1588ee83d19b {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 779.073293] env[61867]: DEBUG nova.compute.manager [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Build of instance 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b was re-scheduled: Binding failed for port 3b1112dd-2d7d-453e-ad9e-677480ae22c4, please check neutron logs for more information. {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 779.073731] env[61867]: DEBUG nova.compute.manager [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Unplugging VIFs for instance {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 779.073958] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "refresh_cache-311a8d4f-68dc-4e0c-acdb-d3929afb9a1b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.074116] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "refresh_cache-311a8d4f-68dc-4e0c-acdb-d3929afb9a1b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.074844] env[61867]: DEBUG nova.network.neutron [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 779.075752] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.082700] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276444, 'name': CreateVM_Task, 'duration_secs': 0.323679} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.082856] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 779.083497] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.083609] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.083915] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 779.084368] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-548c25fc-4999-43c1-8ea4-b05f42c04ee4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.088728] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Waiting for the task: (returnval){ [ 779.088728] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b1078a-5c17-6a8f-9894-f11cce7b5f80" [ 779.088728] env[61867]: _type = "Task" [ 779.088728] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.096298] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b1078a-5c17-6a8f-9894-f11cce7b5f80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.244144] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.257563] env[61867]: DEBUG nova.network.neutron [req-a45a47f0-5bc1-4fa5-8a6c-0e9421713f87 req-23294e6d-4c5d-4771-ae73-2e2eb4780ec0 service nova] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Updated VIF entry in instance network info cache for port fdee8656-849f-4969-8197-bce85b819979. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 779.257949] env[61867]: DEBUG nova.network.neutron [req-a45a47f0-5bc1-4fa5-8a6c-0e9421713f87 req-23294e6d-4c5d-4771-ae73-2e2eb4780ec0 service nova] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Updating instance_info_cache with network_info: [{"id": "fdee8656-849f-4969-8197-bce85b819979", "address": "fa:16:3e:15:81:ae", "network": {"id": "b284d3b5-fef2-4159-a0d1-d44cc9d08c5f", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1659399239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1d8c9b79b3a44e6ac62126ab3ada96b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdee8656-84", "ovs_interfaceid": "fdee8656-849f-4969-8197-bce85b819979", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.291952] env[61867]: INFO nova.compute.manager [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] [instance: 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce] Took 1.02 seconds to deallocate network for instance. [ 779.501191] env[61867]: INFO nova.compute.manager [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Took 39.46 seconds to build instance. [ 779.596042] env[61867]: DEBUG nova.network.neutron [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.603863] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b1078a-5c17-6a8f-9894-f11cce7b5f80, 'name': SearchDatastore_Task, 'duration_secs': 0.009188} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.609222] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.609222] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 779.609222] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.609222] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.609388] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 779.609388] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-75e166fb-c886-493a-93c1-0b78aed502e6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.616884] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 779.616884] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 779.617582] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e583dc37-09c3-406d-8fb0-654a82441502 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.627273] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Waiting for the task: (returnval){ [ 779.627273] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e3741d-8aa8-622c-6bb5-f3375113850b" [ 779.627273] env[61867]: _type = "Task" [ 779.627273] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.638562] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e3741d-8aa8-622c-6bb5-f3375113850b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.680532] env[61867]: DEBUG nova.network.neutron [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.760864] env[61867]: DEBUG oslo_concurrency.lockutils [req-a45a47f0-5bc1-4fa5-8a6c-0e9421713f87 req-23294e6d-4c5d-4771-ae73-2e2eb4780ec0 service nova] Releasing lock "refresh_cache-f0b6273f-4177-44a3-8dd1-f65faeb07539" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.761490] env[61867]: DEBUG nova.compute.manager [req-a45a47f0-5bc1-4fa5-8a6c-0e9421713f87 req-23294e6d-4c5d-4771-ae73-2e2eb4780ec0 service nova] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Received event network-vif-deleted-5661376f-2454-4693-a106-69cb78133962 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 779.761490] env[61867]: DEBUG nova.compute.manager [req-a45a47f0-5bc1-4fa5-8a6c-0e9421713f87 req-23294e6d-4c5d-4771-ae73-2e2eb4780ec0 service nova] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Received event network-vif-deleted-471a814b-d572-4c6b-bdbf-9e02e2091d77 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 779.879892] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-414107a5-3562-4ce4-a5db-0d1da7e5e570 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.889512] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b811039-8115-4e7d-84ff-beba4a811ec7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.936015] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f44610-7862-434b-95d0-831223e9f5bb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.946891] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f25d8765-8aae-4463-a24b-9f2e7bdd4ed7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.967866] env[61867]: DEBUG nova.compute.provider_tree [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.002601] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ee023bf0-eb94-4a68-ae55-f9647413b3de tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Lock "c7203cea-2c2d-4a97-8ff4-3d33cd8443f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.768s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.137703] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e3741d-8aa8-622c-6bb5-f3375113850b, 'name': SearchDatastore_Task, 'duration_secs': 0.014638} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.138540] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff68f4cd-43fd-48c4-aa39-cf459bf92dbf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.144211] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Waiting for the task: (returnval){ [ 780.144211] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52692483-210a-c646-e109-5b6c1a4a94ac" [ 780.144211] env[61867]: _type = "Task" [ 780.144211] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.152150] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52692483-210a-c646-e109-5b6c1a4a94ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.183023] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "refresh_cache-311a8d4f-68dc-4e0c-acdb-d3929afb9a1b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.183270] env[61867]: DEBUG nova.compute.manager [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61867) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 780.183458] env[61867]: DEBUG nova.compute.manager [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 780.183623] env[61867]: DEBUG nova.network.neutron [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 780.200068] env[61867]: DEBUG nova.network.neutron [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.242923] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Acquiring lock "c7203cea-2c2d-4a97-8ff4-3d33cd8443f7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.243209] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Lock "c7203cea-2c2d-4a97-8ff4-3d33cd8443f7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.243414] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Acquiring lock "c7203cea-2c2d-4a97-8ff4-3d33cd8443f7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.243593] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Lock "c7203cea-2c2d-4a97-8ff4-3d33cd8443f7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.243758] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Lock "c7203cea-2c2d-4a97-8ff4-3d33cd8443f7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.245842] env[61867]: INFO nova.compute.manager [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Terminating instance [ 780.247999] env[61867]: DEBUG nova.compute.manager [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 780.248241] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 780.249073] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2da7e55-bf17-4fec-aad5-6d56c23cd5fe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.256937] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 780.257183] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f4ebf62-bcf5-4646-8694-2178e0a53770 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.263663] env[61867]: DEBUG oslo_vmware.api [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Waiting for the task: (returnval){ [ 780.263663] env[61867]: value = "task-1276445" [ 780.263663] env[61867]: _type = "Task" [ 780.263663] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.271110] env[61867]: DEBUG oslo_vmware.api [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': task-1276445, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.323075] env[61867]: INFO nova.scheduler.client.report [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Deleted allocations for instance 8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce [ 780.471366] env[61867]: DEBUG nova.scheduler.client.report [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 780.505465] env[61867]: DEBUG nova.compute.manager [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 780.654783] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52692483-210a-c646-e109-5b6c1a4a94ac, 'name': SearchDatastore_Task, 'duration_secs': 0.039262} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.655066] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.655327] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] f0b6273f-4177-44a3-8dd1-f65faeb07539/f0b6273f-4177-44a3-8dd1-f65faeb07539.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 780.655579] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-607c07b6-efcc-4541-bb08-47c5b6d4bdb3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.661441] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Waiting for the task: (returnval){ [ 780.661441] env[61867]: value = "task-1276446" [ 780.661441] env[61867]: _type = "Task" [ 780.661441] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.668936] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': task-1276446, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.703906] env[61867]: DEBUG nova.network.neutron [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.774116] env[61867]: DEBUG oslo_vmware.api [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': task-1276445, 'name': PowerOffVM_Task, 'duration_secs': 0.157663} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.774416] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 780.774603] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 780.774850] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a095d4b7-5dab-433a-ad36-a096d92850ea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.831124] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bf6821bf-bf4d-4913-bd0b-54745b8e2be1 tempest-ListImageFiltersTestJSON-1347798975 tempest-ListImageFiltersTestJSON-1347798975-project-member] Lock "8cb463ed-d4d2-4ab5-a6f5-5a8df22c07ce" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 178.393s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.834999] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 780.835256] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 780.835440] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Deleting the datastore file [datastore2] c7203cea-2c2d-4a97-8ff4-3d33cd8443f7 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 780.835703] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-60f2afa2-581a-4906-aef8-812af1c1dc16 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.842866] env[61867]: DEBUG oslo_vmware.api [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Waiting for the task: (returnval){ [ 780.842866] env[61867]: value = "task-1276448" [ 780.842866] env[61867]: _type = "Task" [ 780.842866] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.852714] env[61867]: DEBUG oslo_vmware.api [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': task-1276448, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.977153] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.909s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.982340] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.877s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.984030] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.984030] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61867) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 780.984030] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.098s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.985240] env[61867]: INFO nova.compute.claims [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 780.988702] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be657f83-07d3-443b-b98d-87376d6b9ca8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.998174] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d22b44-6262-448f-83be-bbe8688cf795 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.017854] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f921095-183e-4891-bbeb-ccd59b039bee {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.025064] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f298257-7d10-47f2-aa51-e1ea310bb2b2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.031393] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.056030] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181498MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61867) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 781.056030] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.057932] env[61867]: INFO nova.scheduler.client.report [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Deleted allocations for instance 8656c3b8-b7b1-442a-9719-1588ee83d19b [ 781.171217] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': task-1276446, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.43792} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.171470] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] f0b6273f-4177-44a3-8dd1-f65faeb07539/f0b6273f-4177-44a3-8dd1-f65faeb07539.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 781.171677] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 781.171933] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-55a6ef78-41d9-415e-8ba5-73cc926d67f9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.177948] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Waiting for the task: (returnval){ [ 781.177948] env[61867]: value = "task-1276449" [ 781.177948] env[61867]: _type = "Task" [ 781.177948] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.185106] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': task-1276449, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.206969] env[61867]: INFO nova.compute.manager [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b] Took 1.02 seconds to deallocate network for instance. [ 781.333433] env[61867]: DEBUG nova.compute.manager [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 781.355021] env[61867]: DEBUG oslo_vmware.api [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Task: {'id': task-1276448, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.365381} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.357027] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 781.357027] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 781.357027] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 781.357027] env[61867]: INFO nova.compute.manager [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Took 1.11 seconds to destroy the instance on the hypervisor. [ 781.357189] env[61867]: DEBUG oslo.service.loopingcall [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 781.359623] env[61867]: DEBUG nova.compute.manager [-] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 781.359729] env[61867]: DEBUG nova.network.neutron [-] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 781.566655] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b702dab-86b2-4a20-9318-1a4fd1899e91 tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "8656c3b8-b7b1-442a-9719-1588ee83d19b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.449s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.688513] env[61867]: DEBUG nova.compute.manager [req-ed87a18d-6852-4df2-be46-548dfb508ebc req-74213cca-df66-4475-9320-cd941fe2fc79 service nova] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Received event network-vif-deleted-d14dbc28-af77-4bdc-88e1-c56a5c320657 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 781.688718] env[61867]: INFO nova.compute.manager [req-ed87a18d-6852-4df2-be46-548dfb508ebc req-74213cca-df66-4475-9320-cd941fe2fc79 service nova] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Neutron deleted interface d14dbc28-af77-4bdc-88e1-c56a5c320657; detaching it from the instance and deleting it from the info cache [ 781.688888] env[61867]: DEBUG nova.network.neutron [req-ed87a18d-6852-4df2-be46-548dfb508ebc req-74213cca-df66-4475-9320-cd941fe2fc79 service nova] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.695963] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': task-1276449, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067427} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.696167] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 781.696904] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f86d82e-a688-4443-8137-afa8b3ce7737 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.723376] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] f0b6273f-4177-44a3-8dd1-f65faeb07539/f0b6273f-4177-44a3-8dd1-f65faeb07539.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 781.723921] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec13ca95-fc68-46fb-9df3-1b3a6ed45726 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.744218] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Waiting for the task: (returnval){ [ 781.744218] env[61867]: value = "task-1276450" [ 781.744218] env[61867]: _type = "Task" [ 781.744218] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.753523] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': task-1276450, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.853959] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.863695] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "b1ec32c5-0642-4dca-ad43-05a5172d04d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.863877] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "b1ec32c5-0642-4dca-ad43-05a5172d04d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.864125] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "b1ec32c5-0642-4dca-ad43-05a5172d04d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.864316] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "b1ec32c5-0642-4dca-ad43-05a5172d04d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.864483] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "b1ec32c5-0642-4dca-ad43-05a5172d04d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.866646] env[61867]: INFO nova.compute.manager [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Terminating instance [ 781.868332] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "refresh_cache-b1ec32c5-0642-4dca-ad43-05a5172d04d7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.868504] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquired lock "refresh_cache-b1ec32c5-0642-4dca-ad43-05a5172d04d7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.868687] env[61867]: DEBUG nova.network.neutron [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 782.121540] env[61867]: DEBUG nova.network.neutron [-] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.192035] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-36b3c44e-c838-4cdf-b74c-2982ede0ab07 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.205684] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ab215d-4229-42cf-a0a9-35c014cf41a2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.234338] env[61867]: DEBUG nova.compute.manager [req-ed87a18d-6852-4df2-be46-548dfb508ebc req-74213cca-df66-4475-9320-cd941fe2fc79 service nova] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Detach interface failed, port_id=d14dbc28-af77-4bdc-88e1-c56a5c320657, reason: Instance c7203cea-2c2d-4a97-8ff4-3d33cd8443f7 could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 782.239628] env[61867]: INFO nova.scheduler.client.report [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Deleted allocations for instance 311a8d4f-68dc-4e0c-acdb-d3929afb9a1b [ 782.257172] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': task-1276450, 'name': ReconfigVM_Task, 'duration_secs': 0.291403} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.257459] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Reconfigured VM instance instance-00000039 to attach disk [datastore2] f0b6273f-4177-44a3-8dd1-f65faeb07539/f0b6273f-4177-44a3-8dd1-f65faeb07539.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 782.258076] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4bd4a719-c1ce-4636-abd9-c935e9d0bd43 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.264677] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Waiting for the task: (returnval){ [ 782.264677] env[61867]: value = "task-1276451" [ 782.264677] env[61867]: _type = "Task" [ 782.264677] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.274063] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': task-1276451, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.279210] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc42cfd-1a34-4157-b82d-ebb55440b436 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.287412] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad8d59ea-4e98-4bff-bb66-e23518a397c7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.318650] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd26040-9ab3-4848-b64c-b8bda3214212 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.326334] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d96cd6e7-1d04-4c5c-b0bd-7437619c32f8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.339720] env[61867]: DEBUG nova.compute.provider_tree [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.391248] env[61867]: DEBUG nova.network.neutron [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.468399] env[61867]: DEBUG nova.network.neutron [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.625061] env[61867]: INFO nova.compute.manager [-] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Took 1.27 seconds to deallocate network for instance. [ 782.752716] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4715c991-ad5b-4df9-9df3-e45b2ec4af4d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "311a8d4f-68dc-4e0c-acdb-d3929afb9a1b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 178.307s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.774870] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': task-1276451, 'name': Rename_Task, 'duration_secs': 0.216574} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.775168] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 782.775403] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-987f8035-2087-433d-9aaa-fc75a95e60ee {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.781189] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Waiting for the task: (returnval){ [ 782.781189] env[61867]: value = "task-1276452" [ 782.781189] env[61867]: _type = "Task" [ 782.781189] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.790712] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': task-1276452, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.843019] env[61867]: DEBUG nova.scheduler.client.report [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.967195] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Releasing lock "refresh_cache-b1ec32c5-0642-4dca-ad43-05a5172d04d7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.967696] env[61867]: DEBUG nova.compute.manager [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 782.967947] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 782.968837] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ca1742b-310e-449b-949f-2b867a4a1800 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.978787] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 782.979050] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-81c5689e-83a6-43ae-9667-efcab4763598 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.985208] env[61867]: DEBUG oslo_vmware.api [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 782.985208] env[61867]: value = "task-1276453" [ 782.985208] env[61867]: _type = "Task" [ 782.985208] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.995563] env[61867]: DEBUG oslo_vmware.api [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276453, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.132458] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.258720] env[61867]: DEBUG nova.compute.manager [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 783.291688] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': task-1276452, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.351022] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.351022] env[61867]: DEBUG nova.compute.manager [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 783.352516] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.891s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.354103] env[61867]: INFO nova.compute.claims [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.495559] env[61867]: DEBUG oslo_vmware.api [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276453, 'name': PowerOffVM_Task, 'duration_secs': 0.121246} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.495823] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 783.495986] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 783.496354] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e5b324c5-c158-4fec-b77b-a8398065f724 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.519429] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 783.519645] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 783.519993] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Deleting the datastore file [datastore2] b1ec32c5-0642-4dca-ad43-05a5172d04d7 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 783.520166] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b0eb1990-94cb-4cf1-b19b-e3108e152987 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.526118] env[61867]: DEBUG oslo_vmware.api [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for the task: (returnval){ [ 783.526118] env[61867]: value = "task-1276455" [ 783.526118] env[61867]: _type = "Task" [ 783.526118] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.534080] env[61867]: DEBUG oslo_vmware.api [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276455, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.790123] env[61867]: DEBUG oslo_concurrency.lockutils [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.796529] env[61867]: DEBUG oslo_vmware.api [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': task-1276452, 'name': PowerOnVM_Task, 'duration_secs': 0.573782} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.796860] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 783.797186] env[61867]: INFO nova.compute.manager [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Took 7.60 seconds to spawn the instance on the hypervisor. [ 783.797571] env[61867]: DEBUG nova.compute.manager [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 783.798407] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9223d1e-8328-4d38-9ee8-ed0098f85b51 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.861023] env[61867]: DEBUG nova.compute.utils [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 783.864376] env[61867]: DEBUG nova.compute.manager [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 783.864711] env[61867]: DEBUG nova.network.neutron [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 783.956110] env[61867]: DEBUG nova.policy [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae1ff5effba94a5883456b6115e6c5d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f840df7fdbbd499cb6a10bf63ad3c837', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 784.038771] env[61867]: DEBUG oslo_vmware.api [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Task: {'id': task-1276455, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.103431} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.039077] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 784.039505] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 784.039505] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 784.040112] env[61867]: INFO nova.compute.manager [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Took 1.07 seconds to destroy the instance on the hypervisor. [ 784.040312] env[61867]: DEBUG oslo.service.loopingcall [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.040649] env[61867]: DEBUG nova.compute.manager [-] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 784.040743] env[61867]: DEBUG nova.network.neutron [-] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 784.056151] env[61867]: DEBUG nova.network.neutron [-] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.323135] env[61867]: INFO nova.compute.manager [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Took 39.95 seconds to build instance. [ 784.366582] env[61867]: DEBUG nova.compute.manager [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 784.558415] env[61867]: DEBUG nova.network.neutron [-] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.572404] env[61867]: DEBUG nova.network.neutron [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Successfully created port: 32f22e84-5091-47cf-a1be-91303f571008 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 784.710349] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c5a18c-2e73-416b-9589-f8da52febed3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.718349] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb76768-153b-41ed-8b4d-3a744fceada0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.754591] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfdf55f5-89fb-4285-be4b-2406a9d4c6eb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.763487] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef80c7b-d513-4f2c-aa0c-c9431fafff1c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.779980] env[61867]: DEBUG nova.compute.provider_tree [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.827963] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb88d2b9-af69-4dbe-a362-ad76c8f8dc77 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Lock "f0b6273f-4177-44a3-8dd1-f65faeb07539" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.745s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.061200] env[61867]: INFO nova.compute.manager [-] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Took 1.02 seconds to deallocate network for instance. [ 785.284701] env[61867]: DEBUG nova.scheduler.client.report [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.332690] env[61867]: DEBUG nova.compute.manager [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 785.381019] env[61867]: DEBUG nova.compute.manager [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 785.407029] env[61867]: DEBUG nova.virt.hardware [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 785.407283] env[61867]: DEBUG nova.virt.hardware [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 785.407441] env[61867]: DEBUG nova.virt.hardware [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 785.407617] env[61867]: DEBUG nova.virt.hardware [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 785.407758] env[61867]: DEBUG nova.virt.hardware [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 785.407902] env[61867]: DEBUG nova.virt.hardware [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 785.408282] env[61867]: DEBUG nova.virt.hardware [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 785.408437] env[61867]: DEBUG nova.virt.hardware [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 785.408641] env[61867]: DEBUG nova.virt.hardware [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 785.408833] env[61867]: DEBUG nova.virt.hardware [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 785.409069] env[61867]: DEBUG nova.virt.hardware [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 785.409978] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64af2e91-f425-4028-9934-8a2f5de7f225 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.418580] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b0594f-386b-4b37-9cb4-9cdd3c6054e6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.569688] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.772763] env[61867]: DEBUG oslo_concurrency.lockutils [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Acquiring lock "ad699b24-d01d-4d7a-815f-c6b10286012d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.772763] env[61867]: DEBUG oslo_concurrency.lockutils [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Lock "ad699b24-d01d-4d7a-815f-c6b10286012d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.794819] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.442s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.795663] env[61867]: DEBUG nova.compute.manager [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.800037] env[61867]: DEBUG oslo_concurrency.lockutils [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.012s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.800692] env[61867]: DEBUG nova.objects.instance [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Lazy-loading 'resources' on Instance uuid 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 785.866674] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.879795] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.879928] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.081808] env[61867]: DEBUG nova.compute.manager [req-92e972a5-2673-4205-83d2-8938eba41e6d req-fc3d697d-d1df-4901-b608-f61fdc6df375 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Received event network-vif-plugged-32f22e84-5091-47cf-a1be-91303f571008 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 786.082043] env[61867]: DEBUG oslo_concurrency.lockutils [req-92e972a5-2673-4205-83d2-8938eba41e6d req-fc3d697d-d1df-4901-b608-f61fdc6df375 service nova] Acquiring lock "905ee661-d5bd-4121-9bd0-fe68bdabd2fd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.082243] env[61867]: DEBUG oslo_concurrency.lockutils [req-92e972a5-2673-4205-83d2-8938eba41e6d req-fc3d697d-d1df-4901-b608-f61fdc6df375 service nova] Lock "905ee661-d5bd-4121-9bd0-fe68bdabd2fd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.082412] env[61867]: DEBUG oslo_concurrency.lockutils [req-92e972a5-2673-4205-83d2-8938eba41e6d req-fc3d697d-d1df-4901-b608-f61fdc6df375 service nova] Lock "905ee661-d5bd-4121-9bd0-fe68bdabd2fd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.082575] env[61867]: DEBUG nova.compute.manager [req-92e972a5-2673-4205-83d2-8938eba41e6d req-fc3d697d-d1df-4901-b608-f61fdc6df375 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] No waiting events found dispatching network-vif-plugged-32f22e84-5091-47cf-a1be-91303f571008 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 786.082736] env[61867]: WARNING nova.compute.manager [req-92e972a5-2673-4205-83d2-8938eba41e6d req-fc3d697d-d1df-4901-b608-f61fdc6df375 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Received unexpected event network-vif-plugged-32f22e84-5091-47cf-a1be-91303f571008 for instance with vm_state building and task_state spawning. [ 786.098445] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Acquiring lock "f0b6273f-4177-44a3-8dd1-f65faeb07539" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.099037] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Lock "f0b6273f-4177-44a3-8dd1-f65faeb07539" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.099037] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Acquiring lock "f0b6273f-4177-44a3-8dd1-f65faeb07539-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.099161] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Lock "f0b6273f-4177-44a3-8dd1-f65faeb07539-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.099249] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Lock "f0b6273f-4177-44a3-8dd1-f65faeb07539-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.102072] env[61867]: INFO nova.compute.manager [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Terminating instance [ 786.104318] env[61867]: DEBUG nova.compute.manager [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 786.104509] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 786.105338] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b1697c-18dc-427e-8c0d-25c41ab00a0e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.113235] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 786.113535] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-69d81ae9-82b7-4814-8298-eb8b47e5c52d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.123545] env[61867]: DEBUG oslo_vmware.api [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Waiting for the task: (returnval){ [ 786.123545] env[61867]: value = "task-1276456" [ 786.123545] env[61867]: _type = "Task" [ 786.123545] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.133741] env[61867]: DEBUG oslo_vmware.api [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': task-1276456, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.182514] env[61867]: DEBUG nova.network.neutron [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Successfully updated port: 32f22e84-5091-47cf-a1be-91303f571008 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 786.302096] env[61867]: DEBUG nova.compute.utils [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 786.303973] env[61867]: DEBUG nova.compute.manager [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 786.304749] env[61867]: DEBUG nova.network.neutron [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 786.376570] env[61867]: DEBUG nova.policy [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a013ca682054d16a6702230a22c180f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c28a8e133444440eb83dc6a848ef591d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 786.635070] env[61867]: DEBUG oslo_vmware.api [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': task-1276456, 'name': PowerOffVM_Task, 'duration_secs': 0.216826} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.635354] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 786.635493] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 786.635733] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2d9cb185-d33c-48c5-b820-dd62d182d2c3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.651291] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85aefc95-e372-41be-858c-54b449ad2835 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.659164] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0346eaff-612f-47ca-ba76-1ae1465ba27f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.697009] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquiring lock "refresh_cache-905ee661-d5bd-4121-9bd0-fe68bdabd2fd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.697327] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquired lock "refresh_cache-905ee661-d5bd-4121-9bd0-fe68bdabd2fd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.697579] env[61867]: DEBUG nova.network.neutron [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 786.701997] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba56843-4814-461c-8949-a0fe7d262bdc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.706136] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 786.706463] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 786.706747] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Deleting the datastore file [datastore2] f0b6273f-4177-44a3-8dd1-f65faeb07539 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 786.710872] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-25689592-2928-4509-83be-e8f5dbd73aea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.725134] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f28e8423-a653-4655-9d24-aa6b529cf1e6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.731476] env[61867]: DEBUG oslo_vmware.api [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Waiting for the task: (returnval){ [ 786.731476] env[61867]: value = "task-1276458" [ 786.731476] env[61867]: _type = "Task" [ 786.731476] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.749778] env[61867]: DEBUG nova.compute.provider_tree [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.755806] env[61867]: DEBUG oslo_vmware.api [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': task-1276458, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.763589] env[61867]: DEBUG nova.network.neutron [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 786.812870] env[61867]: DEBUG nova.compute.manager [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 786.849557] env[61867]: DEBUG nova.network.neutron [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Successfully created port: 8546f778-9a66-4428-a784-38b85f0f3f98 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 787.071444] env[61867]: DEBUG nova.network.neutron [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Updating instance_info_cache with network_info: [{"id": "32f22e84-5091-47cf-a1be-91303f571008", "address": "fa:16:3e:ad:02:08", "network": {"id": "55ecbbe4-0b61-4064-8187-e30c1e444949", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-176665084-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f840df7fdbbd499cb6a10bf63ad3c837", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbdab640-5fea-4254-8bd3-f855b7eaca0d", "external-id": "nsx-vlan-transportzone-615", "segmentation_id": 615, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32f22e84-50", "ovs_interfaceid": "32f22e84-5091-47cf-a1be-91303f571008", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.242742] env[61867]: DEBUG oslo_vmware.api [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Task: {'id': task-1276458, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.365441} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.243106] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 787.243310] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 787.243488] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 787.243662] env[61867]: INFO nova.compute.manager [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Took 1.14 seconds to destroy the instance on the hypervisor. [ 787.243919] env[61867]: DEBUG oslo.service.loopingcall [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.244165] env[61867]: DEBUG nova.compute.manager [-] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 787.244263] env[61867]: DEBUG nova.network.neutron [-] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 787.253159] env[61867]: DEBUG nova.scheduler.client.report [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.574306] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Releasing lock "refresh_cache-905ee661-d5bd-4121-9bd0-fe68bdabd2fd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.574739] env[61867]: DEBUG nova.compute.manager [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Instance network_info: |[{"id": "32f22e84-5091-47cf-a1be-91303f571008", "address": "fa:16:3e:ad:02:08", "network": {"id": "55ecbbe4-0b61-4064-8187-e30c1e444949", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-176665084-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f840df7fdbbd499cb6a10bf63ad3c837", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbdab640-5fea-4254-8bd3-f855b7eaca0d", "external-id": "nsx-vlan-transportzone-615", "segmentation_id": 615, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32f22e84-50", "ovs_interfaceid": "32f22e84-5091-47cf-a1be-91303f571008", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 787.575178] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ad:02:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dbdab640-5fea-4254-8bd3-f855b7eaca0d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '32f22e84-5091-47cf-a1be-91303f571008', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 787.583392] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Creating folder: Project (f840df7fdbbd499cb6a10bf63ad3c837). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 787.585142] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b835cb14-69f7-43fd-94e5-fedde422018c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.587844] env[61867]: DEBUG nova.compute.manager [req-1da12d1f-f184-4584-9139-ba8deb933ed6 req-632ee53d-7564-4573-af63-071a5272c348 service nova] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Received event network-vif-deleted-fdee8656-849f-4969-8197-bce85b819979 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.588042] env[61867]: INFO nova.compute.manager [req-1da12d1f-f184-4584-9139-ba8deb933ed6 req-632ee53d-7564-4573-af63-071a5272c348 service nova] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Neutron deleted interface fdee8656-849f-4969-8197-bce85b819979; detaching it from the instance and deleting it from the info cache [ 787.588219] env[61867]: DEBUG nova.network.neutron [req-1da12d1f-f184-4584-9139-ba8deb933ed6 req-632ee53d-7564-4573-af63-071a5272c348 service nova] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.598625] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Created folder: Project (f840df7fdbbd499cb6a10bf63ad3c837) in parent group-v274258. [ 787.598625] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Creating folder: Instances. Parent ref: group-v274300. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 787.598727] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-45e65451-431a-42ca-9f62-35b7be34bf17 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.611040] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Created folder: Instances in parent group-v274300. [ 787.611285] env[61867]: DEBUG oslo.service.loopingcall [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.611469] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 787.611672] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-227efe30-7e3a-4ab1-875a-d27fc99d672a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.631311] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 787.631311] env[61867]: value = "task-1276461" [ 787.631311] env[61867]: _type = "Task" [ 787.631311] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.642824] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276461, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.758300] env[61867]: DEBUG oslo_concurrency.lockutils [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.958s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.761764] env[61867]: DEBUG oslo_concurrency.lockutils [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.797s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.762852] env[61867]: INFO nova.compute.claims [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 787.781827] env[61867]: INFO nova.scheduler.client.report [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Deleted allocations for instance 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef [ 787.823731] env[61867]: DEBUG nova.compute.manager [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 787.848455] env[61867]: DEBUG nova.virt.hardware [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 787.848681] env[61867]: DEBUG nova.virt.hardware [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 787.848839] env[61867]: DEBUG nova.virt.hardware [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.849028] env[61867]: DEBUG nova.virt.hardware [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 787.849345] env[61867]: DEBUG nova.virt.hardware [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.849345] env[61867]: DEBUG nova.virt.hardware [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 787.849518] env[61867]: DEBUG nova.virt.hardware [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 787.849772] env[61867]: DEBUG nova.virt.hardware [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 787.850038] env[61867]: DEBUG nova.virt.hardware [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 787.850150] env[61867]: DEBUG nova.virt.hardware [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 787.850325] env[61867]: DEBUG nova.virt.hardware [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 787.851511] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eea8cc4-56f2-4e31-812a-19ebfb60e21e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.860496] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dbc2d42-0880-4826-bccd-f91352fcff9c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.000575] env[61867]: DEBUG nova.network.neutron [-] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.092127] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-52f4baf3-9cfe-434c-9873-7dd5917cae76 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.101324] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc58b334-8f66-4861-9faf-1b2856535335 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.113875] env[61867]: DEBUG nova.compute.manager [req-15d5ddee-1858-45c0-9bef-8cf0778d5ed7 req-74de2282-d07b-4f1d-b8ea-22231ca6dae2 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Received event network-changed-32f22e84-5091-47cf-a1be-91303f571008 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.114074] env[61867]: DEBUG nova.compute.manager [req-15d5ddee-1858-45c0-9bef-8cf0778d5ed7 req-74de2282-d07b-4f1d-b8ea-22231ca6dae2 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Refreshing instance network info cache due to event network-changed-32f22e84-5091-47cf-a1be-91303f571008. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 788.114285] env[61867]: DEBUG oslo_concurrency.lockutils [req-15d5ddee-1858-45c0-9bef-8cf0778d5ed7 req-74de2282-d07b-4f1d-b8ea-22231ca6dae2 service nova] Acquiring lock "refresh_cache-905ee661-d5bd-4121-9bd0-fe68bdabd2fd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.114577] env[61867]: DEBUG oslo_concurrency.lockutils [req-15d5ddee-1858-45c0-9bef-8cf0778d5ed7 req-74de2282-d07b-4f1d-b8ea-22231ca6dae2 service nova] Acquired lock "refresh_cache-905ee661-d5bd-4121-9bd0-fe68bdabd2fd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.115580] env[61867]: DEBUG nova.network.neutron [req-15d5ddee-1858-45c0-9bef-8cf0778d5ed7 req-74de2282-d07b-4f1d-b8ea-22231ca6dae2 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Refreshing network info cache for port 32f22e84-5091-47cf-a1be-91303f571008 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 788.128391] env[61867]: DEBUG nova.compute.manager [req-1da12d1f-f184-4584-9139-ba8deb933ed6 req-632ee53d-7564-4573-af63-071a5272c348 service nova] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Detach interface failed, port_id=fdee8656-849f-4969-8197-bce85b819979, reason: Instance f0b6273f-4177-44a3-8dd1-f65faeb07539 could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 788.140036] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276461, 'name': CreateVM_Task, 'duration_secs': 0.341522} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.140133] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 788.140740] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.140899] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.141227] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 788.141469] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09050dbc-2406-4fdb-86b6-add2d255f11f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.145729] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Waiting for the task: (returnval){ [ 788.145729] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5245d0cf-2a76-5874-d6e2-fc35acc32cdd" [ 788.145729] env[61867]: _type = "Task" [ 788.145729] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.153175] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5245d0cf-2a76-5874-d6e2-fc35acc32cdd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.290018] env[61867]: DEBUG oslo_concurrency.lockutils [None req-673a0ab3-1fde-4690-b2b1-9d1927f35d6f tempest-ServerAddressesNegativeTestJSON-341398244 tempest-ServerAddressesNegativeTestJSON-341398244-project-member] Lock "7c0badee-0e2b-4a13-89b1-e57df5c2c9ef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.435s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.362244] env[61867]: DEBUG nova.network.neutron [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Successfully updated port: 8546f778-9a66-4428-a784-38b85f0f3f98 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 788.503803] env[61867]: INFO nova.compute.manager [-] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Took 1.26 seconds to deallocate network for instance. [ 788.663327] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5245d0cf-2a76-5874-d6e2-fc35acc32cdd, 'name': SearchDatastore_Task, 'duration_secs': 0.009661} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.663327] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.663327] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 788.663327] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.663651] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.663651] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 788.663651] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be01fb93-0819-4928-8f0b-5275ce86779d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.674017] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 788.674314] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 788.675036] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eac22df5-a53e-46d7-9ee0-24efb6148f38 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.679826] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Waiting for the task: (returnval){ [ 788.679826] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]524fdc8d-17b6-92de-4111-47ee6b26f278" [ 788.679826] env[61867]: _type = "Task" [ 788.679826] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.687659] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]524fdc8d-17b6-92de-4111-47ee6b26f278, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.834365] env[61867]: DEBUG nova.network.neutron [req-15d5ddee-1858-45c0-9bef-8cf0778d5ed7 req-74de2282-d07b-4f1d-b8ea-22231ca6dae2 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Updated VIF entry in instance network info cache for port 32f22e84-5091-47cf-a1be-91303f571008. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 788.834715] env[61867]: DEBUG nova.network.neutron [req-15d5ddee-1858-45c0-9bef-8cf0778d5ed7 req-74de2282-d07b-4f1d-b8ea-22231ca6dae2 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Updating instance_info_cache with network_info: [{"id": "32f22e84-5091-47cf-a1be-91303f571008", "address": "fa:16:3e:ad:02:08", "network": {"id": "55ecbbe4-0b61-4064-8187-e30c1e444949", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-176665084-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f840df7fdbbd499cb6a10bf63ad3c837", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbdab640-5fea-4254-8bd3-f855b7eaca0d", "external-id": "nsx-vlan-transportzone-615", "segmentation_id": 615, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32f22e84-50", "ovs_interfaceid": "32f22e84-5091-47cf-a1be-91303f571008", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.866342] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "refresh_cache-3022ea79-7c27-40a4-80d6-1a86b354917a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.866572] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "refresh_cache-3022ea79-7c27-40a4-80d6-1a86b354917a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.866772] env[61867]: DEBUG nova.network.neutron [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 789.010248] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.161499] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c9fe04-0fcf-4f0f-99cb-197d5fabf83f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.172025] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f9b51a-5fe2-4654-a76e-0f71f7c5e6d1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.206975] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2b01bfa-9a2e-40d5-bc2a-75c58242c11c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.217028] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]524fdc8d-17b6-92de-4111-47ee6b26f278, 'name': SearchDatastore_Task, 'duration_secs': 0.007868} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.217479] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db9793a6-34ce-4b67-94cf-b9f2ee88888e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.220566] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec67754-6342-468e-bc88-07b544fae0c8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.229516] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Waiting for the task: (returnval){ [ 789.229516] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523eed92-5866-042e-6291-b78953ef9f2d" [ 789.229516] env[61867]: _type = "Task" [ 789.229516] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.237084] env[61867]: DEBUG nova.compute.provider_tree [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.246509] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523eed92-5866-042e-6291-b78953ef9f2d, 'name': SearchDatastore_Task, 'duration_secs': 0.009824} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.246766] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.247029] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 905ee661-d5bd-4121-9bd0-fe68bdabd2fd/905ee661-d5bd-4121-9bd0-fe68bdabd2fd.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 789.247285] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-84f80593-7e29-4286-8059-7c4efcf07124 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.254678] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Waiting for the task: (returnval){ [ 789.254678] env[61867]: value = "task-1276462" [ 789.254678] env[61867]: _type = "Task" [ 789.254678] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.263512] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276462, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.338356] env[61867]: DEBUG oslo_concurrency.lockutils [req-15d5ddee-1858-45c0-9bef-8cf0778d5ed7 req-74de2282-d07b-4f1d-b8ea-22231ca6dae2 service nova] Releasing lock "refresh_cache-905ee661-d5bd-4121-9bd0-fe68bdabd2fd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.401859] env[61867]: DEBUG nova.network.neutron [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.553387] env[61867]: DEBUG nova.network.neutron [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Updating instance_info_cache with network_info: [{"id": "8546f778-9a66-4428-a784-38b85f0f3f98", "address": "fa:16:3e:13:f6:2f", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8546f778-9a", "ovs_interfaceid": "8546f778-9a66-4428-a784-38b85f0f3f98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.742463] env[61867]: DEBUG nova.scheduler.client.report [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 789.766034] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276462, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457566} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.766364] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 905ee661-d5bd-4121-9bd0-fe68bdabd2fd/905ee661-d5bd-4121-9bd0-fe68bdabd2fd.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 789.766619] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 789.766901] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8db2dbd6-8828-455d-b770-d7c5aca82f57 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.773328] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Waiting for the task: (returnval){ [ 789.773328] env[61867]: value = "task-1276463" [ 789.773328] env[61867]: _type = "Task" [ 789.773328] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.781968] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276463, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.056796] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "refresh_cache-3022ea79-7c27-40a4-80d6-1a86b354917a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.057075] env[61867]: DEBUG nova.compute.manager [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Instance network_info: |[{"id": "8546f778-9a66-4428-a784-38b85f0f3f98", "address": "fa:16:3e:13:f6:2f", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8546f778-9a", "ovs_interfaceid": "8546f778-9a66-4428-a784-38b85f0f3f98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 790.057508] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:f6:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74e6f6e0-95e6-4531-99e9-0e78350fb655', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8546f778-9a66-4428-a784-38b85f0f3f98', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 790.065686] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Creating folder: Project (c28a8e133444440eb83dc6a848ef591d). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 790.065790] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fc74c13a-e4da-481c-b430-07a7ab327e51 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.075616] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Created folder: Project (c28a8e133444440eb83dc6a848ef591d) in parent group-v274258. [ 790.075810] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Creating folder: Instances. Parent ref: group-v274303. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 790.076116] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0fe68166-b7f4-4767-aa06-3f9f42e2ec5a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.085909] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Created folder: Instances in parent group-v274303. [ 790.086159] env[61867]: DEBUG oslo.service.loopingcall [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.086348] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 790.086653] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3d724ae1-882c-4a25-9277-a5bf73c8d062 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.105727] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 790.105727] env[61867]: value = "task-1276466" [ 790.105727] env[61867]: _type = "Task" [ 790.105727] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.115075] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276466, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.178656] env[61867]: DEBUG nova.compute.manager [req-58093cd9-9d04-49fd-a2ec-b60a910cbad8 req-b85f8b1b-db26-48b1-bb81-ad79934a697e service nova] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Received event network-vif-plugged-8546f778-9a66-4428-a784-38b85f0f3f98 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.178656] env[61867]: DEBUG oslo_concurrency.lockutils [req-58093cd9-9d04-49fd-a2ec-b60a910cbad8 req-b85f8b1b-db26-48b1-bb81-ad79934a697e service nova] Acquiring lock "3022ea79-7c27-40a4-80d6-1a86b354917a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.178656] env[61867]: DEBUG oslo_concurrency.lockutils [req-58093cd9-9d04-49fd-a2ec-b60a910cbad8 req-b85f8b1b-db26-48b1-bb81-ad79934a697e service nova] Lock "3022ea79-7c27-40a4-80d6-1a86b354917a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.178656] env[61867]: DEBUG oslo_concurrency.lockutils [req-58093cd9-9d04-49fd-a2ec-b60a910cbad8 req-b85f8b1b-db26-48b1-bb81-ad79934a697e service nova] Lock "3022ea79-7c27-40a4-80d6-1a86b354917a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.178656] env[61867]: DEBUG nova.compute.manager [req-58093cd9-9d04-49fd-a2ec-b60a910cbad8 req-b85f8b1b-db26-48b1-bb81-ad79934a697e service nova] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] No waiting events found dispatching network-vif-plugged-8546f778-9a66-4428-a784-38b85f0f3f98 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 790.178801] env[61867]: WARNING nova.compute.manager [req-58093cd9-9d04-49fd-a2ec-b60a910cbad8 req-b85f8b1b-db26-48b1-bb81-ad79934a697e service nova] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Received unexpected event network-vif-plugged-8546f778-9a66-4428-a784-38b85f0f3f98 for instance with vm_state building and task_state spawning. [ 790.178832] env[61867]: DEBUG nova.compute.manager [req-58093cd9-9d04-49fd-a2ec-b60a910cbad8 req-b85f8b1b-db26-48b1-bb81-ad79934a697e service nova] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Received event network-changed-8546f778-9a66-4428-a784-38b85f0f3f98 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.178964] env[61867]: DEBUG nova.compute.manager [req-58093cd9-9d04-49fd-a2ec-b60a910cbad8 req-b85f8b1b-db26-48b1-bb81-ad79934a697e service nova] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Refreshing instance network info cache due to event network-changed-8546f778-9a66-4428-a784-38b85f0f3f98. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 790.179277] env[61867]: DEBUG oslo_concurrency.lockutils [req-58093cd9-9d04-49fd-a2ec-b60a910cbad8 req-b85f8b1b-db26-48b1-bb81-ad79934a697e service nova] Acquiring lock "refresh_cache-3022ea79-7c27-40a4-80d6-1a86b354917a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.179277] env[61867]: DEBUG oslo_concurrency.lockutils [req-58093cd9-9d04-49fd-a2ec-b60a910cbad8 req-b85f8b1b-db26-48b1-bb81-ad79934a697e service nova] Acquired lock "refresh_cache-3022ea79-7c27-40a4-80d6-1a86b354917a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.179477] env[61867]: DEBUG nova.network.neutron [req-58093cd9-9d04-49fd-a2ec-b60a910cbad8 req-b85f8b1b-db26-48b1-bb81-ad79934a697e service nova] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Refreshing network info cache for port 8546f778-9a66-4428-a784-38b85f0f3f98 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 790.249756] env[61867]: DEBUG oslo_concurrency.lockutils [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.488s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.250346] env[61867]: DEBUG nova.compute.manager [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 790.254320] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.967s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.254756] env[61867]: DEBUG nova.objects.instance [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Lazy-loading 'resources' on Instance uuid d0db58a5-d773-4e45-a238-07cfc156bdd2 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 790.284015] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276463, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061383} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.284303] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 790.285139] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07dd099-d36a-442b-9249-52f16728a628 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.307881] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 905ee661-d5bd-4121-9bd0-fe68bdabd2fd/905ee661-d5bd-4121-9bd0-fe68bdabd2fd.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 790.308443] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46ebe89c-f8c7-4a9f-a1df-a358bcf7aecb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.327948] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Waiting for the task: (returnval){ [ 790.327948] env[61867]: value = "task-1276467" [ 790.327948] env[61867]: _type = "Task" [ 790.327948] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.337377] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276467, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.615519] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276466, 'name': CreateVM_Task, 'duration_secs': 0.488945} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.615676] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 790.616373] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.616520] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.616840] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 790.617166] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcdd693a-37b1-4f36-a2d6-6cf0dd70415e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.622026] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 790.622026] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52726a6b-e820-a689-c3be-bbdbeb3faff5" [ 790.622026] env[61867]: _type = "Task" [ 790.622026] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.628935] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52726a6b-e820-a689-c3be-bbdbeb3faff5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.763701] env[61867]: DEBUG nova.compute.utils [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 790.767596] env[61867]: DEBUG nova.compute.manager [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 790.767770] env[61867]: DEBUG nova.network.neutron [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 790.836165] env[61867]: DEBUG nova.policy [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'adcd9eb75ecc4eccb335ebfec207a900', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7316e4f263a9432ab2f9f91484d62f58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 790.844794] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276467, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.914556] env[61867]: DEBUG nova.network.neutron [req-58093cd9-9d04-49fd-a2ec-b60a910cbad8 req-b85f8b1b-db26-48b1-bb81-ad79934a697e service nova] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Updated VIF entry in instance network info cache for port 8546f778-9a66-4428-a784-38b85f0f3f98. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 790.915037] env[61867]: DEBUG nova.network.neutron [req-58093cd9-9d04-49fd-a2ec-b60a910cbad8 req-b85f8b1b-db26-48b1-bb81-ad79934a697e service nova] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Updating instance_info_cache with network_info: [{"id": "8546f778-9a66-4428-a784-38b85f0f3f98", "address": "fa:16:3e:13:f6:2f", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8546f778-9a", "ovs_interfaceid": "8546f778-9a66-4428-a784-38b85f0f3f98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.075128] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a0a1cc-2bf9-42f2-913f-0ac9e3b17b3b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.082720] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8419c515-086a-4c16-a7bc-b789a29a3a6b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.111841] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e286cc28-1d87-436b-b752-275e7c632281 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.118920] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc548ccc-b691-4d9e-a622-273cff8f2a7b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.137368] env[61867]: DEBUG nova.compute.provider_tree [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.141744] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52726a6b-e820-a689-c3be-bbdbeb3faff5, 'name': SearchDatastore_Task, 'duration_secs': 0.008502} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.142980] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.142980] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 791.142980] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.142980] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.143194] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 791.143194] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-13bdf995-c728-4f86-940a-2cf33d1be127 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.151675] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 791.151675] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 791.152423] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-edc54c39-2c8d-48c0-a3d1-346c5198f486 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.157424] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 791.157424] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526e5379-fd7f-63db-6f2b-21f7ca1c26ba" [ 791.157424] env[61867]: _type = "Task" [ 791.157424] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.164980] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526e5379-fd7f-63db-6f2b-21f7ca1c26ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.268218] env[61867]: DEBUG nova.compute.manager [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 791.276119] env[61867]: DEBUG nova.network.neutron [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Successfully created port: d9ab90ee-f219-4de7-af2a-3db270f0f2c6 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 791.339210] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276467, 'name': ReconfigVM_Task, 'duration_secs': 0.749897} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.339493] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 905ee661-d5bd-4121-9bd0-fe68bdabd2fd/905ee661-d5bd-4121-9bd0-fe68bdabd2fd.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 791.340124] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-909e6f90-73d9-4d26-a132-1652f0eb88db {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.345926] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Waiting for the task: (returnval){ [ 791.345926] env[61867]: value = "task-1276468" [ 791.345926] env[61867]: _type = "Task" [ 791.345926] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.355342] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276468, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.418357] env[61867]: DEBUG oslo_concurrency.lockutils [req-58093cd9-9d04-49fd-a2ec-b60a910cbad8 req-b85f8b1b-db26-48b1-bb81-ad79934a697e service nova] Releasing lock "refresh_cache-3022ea79-7c27-40a4-80d6-1a86b354917a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.643754] env[61867]: DEBUG nova.scheduler.client.report [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 791.667479] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526e5379-fd7f-63db-6f2b-21f7ca1c26ba, 'name': SearchDatastore_Task, 'duration_secs': 0.008367} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.668233] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b74eadd7-2b78-492e-bde4-c78ce54213fc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.672881] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 791.672881] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]524fa512-4935-abb2-b9ef-162156b57fe4" [ 791.672881] env[61867]: _type = "Task" [ 791.672881] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.679647] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]524fa512-4935-abb2-b9ef-162156b57fe4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.855183] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276468, 'name': Rename_Task, 'duration_secs': 0.134273} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.855450] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 791.855601] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-45240714-54ca-4b65-b0c5-fce7f5c34847 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.862260] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Waiting for the task: (returnval){ [ 791.862260] env[61867]: value = "task-1276469" [ 791.862260] env[61867]: _type = "Task" [ 791.862260] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.869489] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276469, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.148673] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.894s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.151219] env[61867]: DEBUG oslo_concurrency.lockutils [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.449s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.153029] env[61867]: INFO nova.compute.claims [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 792.174861] env[61867]: INFO nova.scheduler.client.report [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Deleted allocations for instance d0db58a5-d773-4e45-a238-07cfc156bdd2 [ 792.188846] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]524fa512-4935-abb2-b9ef-162156b57fe4, 'name': SearchDatastore_Task, 'duration_secs': 0.008482} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.188846] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.188846] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 3022ea79-7c27-40a4-80d6-1a86b354917a/3022ea79-7c27-40a4-80d6-1a86b354917a.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 792.189020] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36f62f5b-3bd0-4730-a862-ed067d968167 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.196252] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 792.196252] env[61867]: value = "task-1276470" [ 792.196252] env[61867]: _type = "Task" [ 792.196252] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.205094] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276470, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.278643] env[61867]: DEBUG nova.compute.manager [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 792.304510] env[61867]: DEBUG nova.virt.hardware [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 792.304808] env[61867]: DEBUG nova.virt.hardware [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 792.305079] env[61867]: DEBUG nova.virt.hardware [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 792.305325] env[61867]: DEBUG nova.virt.hardware [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 792.305520] env[61867]: DEBUG nova.virt.hardware [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 792.305711] env[61867]: DEBUG nova.virt.hardware [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 792.305965] env[61867]: DEBUG nova.virt.hardware [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 792.306180] env[61867]: DEBUG nova.virt.hardware [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 792.306392] env[61867]: DEBUG nova.virt.hardware [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 792.306598] env[61867]: DEBUG nova.virt.hardware [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 792.306809] env[61867]: DEBUG nova.virt.hardware [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 792.307755] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2448b04-3c17-4aa2-b318-bf05bfef83b4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.315989] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-710c8839-57ca-46f9-b14a-4c48668953eb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.372874] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276469, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.687024] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fa232d40-dd01-465a-afb5-1f7bf62b03cc tempest-ServersAaction247Test-1467043704 tempest-ServersAaction247Test-1467043704-project-member] Lock "d0db58a5-d773-4e45-a238-07cfc156bdd2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.150s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.706862] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276470, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.429983} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.707130] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 3022ea79-7c27-40a4-80d6-1a86b354917a/3022ea79-7c27-40a4-80d6-1a86b354917a.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 792.707345] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 792.707594] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6034b5ef-1f93-49e8-b4bd-5bbbd805b771 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.713850] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 792.713850] env[61867]: value = "task-1276471" [ 792.713850] env[61867]: _type = "Task" [ 792.713850] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.722800] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276471, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.877999] env[61867]: DEBUG oslo_vmware.api [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276469, 'name': PowerOnVM_Task, 'duration_secs': 1.009626} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.878362] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 792.878511] env[61867]: INFO nova.compute.manager [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Took 7.50 seconds to spawn the instance on the hypervisor. [ 792.878691] env[61867]: DEBUG nova.compute.manager [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 792.879536] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3394984f-d0ff-45f1-b7f0-e47f055675d6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.120344] env[61867]: DEBUG nova.compute.manager [req-0ab06d58-6a8c-4164-9828-35ad0872b028 req-ff1e7493-3be5-41bc-addf-6c604140a714 service nova] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Received event network-vif-plugged-d9ab90ee-f219-4de7-af2a-3db270f0f2c6 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.120344] env[61867]: DEBUG oslo_concurrency.lockutils [req-0ab06d58-6a8c-4164-9828-35ad0872b028 req-ff1e7493-3be5-41bc-addf-6c604140a714 service nova] Acquiring lock "be825f62-cb56-4647-be1c-1605536fdc89-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.120344] env[61867]: DEBUG oslo_concurrency.lockutils [req-0ab06d58-6a8c-4164-9828-35ad0872b028 req-ff1e7493-3be5-41bc-addf-6c604140a714 service nova] Lock "be825f62-cb56-4647-be1c-1605536fdc89-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.120344] env[61867]: DEBUG oslo_concurrency.lockutils [req-0ab06d58-6a8c-4164-9828-35ad0872b028 req-ff1e7493-3be5-41bc-addf-6c604140a714 service nova] Lock "be825f62-cb56-4647-be1c-1605536fdc89-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.120344] env[61867]: DEBUG nova.compute.manager [req-0ab06d58-6a8c-4164-9828-35ad0872b028 req-ff1e7493-3be5-41bc-addf-6c604140a714 service nova] [instance: be825f62-cb56-4647-be1c-1605536fdc89] No waiting events found dispatching network-vif-plugged-d9ab90ee-f219-4de7-af2a-3db270f0f2c6 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 793.120731] env[61867]: WARNING nova.compute.manager [req-0ab06d58-6a8c-4164-9828-35ad0872b028 req-ff1e7493-3be5-41bc-addf-6c604140a714 service nova] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Received unexpected event network-vif-plugged-d9ab90ee-f219-4de7-af2a-3db270f0f2c6 for instance with vm_state building and task_state spawning. [ 793.225899] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276471, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058226} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.227150] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 793.227150] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d67187e-8f68-417b-95d0-8743e8fceedd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.248923] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 3022ea79-7c27-40a4-80d6-1a86b354917a/3022ea79-7c27-40a4-80d6-1a86b354917a.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 793.252148] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2309cbf0-ff7a-4db7-9bc2-4a61b0e98d2a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.271359] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 793.271359] env[61867]: value = "task-1276472" [ 793.271359] env[61867]: _type = "Task" [ 793.271359] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.284046] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276472, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.397518] env[61867]: INFO nova.compute.manager [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Took 24.53 seconds to build instance. [ 793.452949] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4662d628-e3d5-4cf3-a0e7-41afb7428e65 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.459873] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212c26aa-ac82-4795-92b5-d4b57d16c20b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.488761] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a1b149-d22e-4ea8-a159-4c8afa597cd9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.496639] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c5970a-4bad-4fe6-a56f-b6b2223c9422 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.510816] env[61867]: DEBUG nova.compute.provider_tree [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 793.700311] env[61867]: DEBUG nova.network.neutron [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Successfully updated port: d9ab90ee-f219-4de7-af2a-3db270f0f2c6 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 793.780700] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276472, 'name': ReconfigVM_Task, 'duration_secs': 0.27914} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.781114] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 3022ea79-7c27-40a4-80d6-1a86b354917a/3022ea79-7c27-40a4-80d6-1a86b354917a.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.781769] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cd6f9b4c-bde2-4053-a519-a0dc9d36daf4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.787845] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 793.787845] env[61867]: value = "task-1276473" [ 793.787845] env[61867]: _type = "Task" [ 793.787845] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.795016] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276473, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.898478] env[61867]: DEBUG nova.compute.manager [req-aaa30d77-1336-4be3-9c1b-cee034e321e4 req-cd80ddeb-c73c-4a2a-a90b-37b33c28d385 service nova] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Received event network-changed-d9ab90ee-f219-4de7-af2a-3db270f0f2c6 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.898724] env[61867]: DEBUG nova.compute.manager [req-aaa30d77-1336-4be3-9c1b-cee034e321e4 req-cd80ddeb-c73c-4a2a-a90b-37b33c28d385 service nova] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Refreshing instance network info cache due to event network-changed-d9ab90ee-f219-4de7-af2a-3db270f0f2c6. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 793.898881] env[61867]: DEBUG oslo_concurrency.lockutils [req-aaa30d77-1336-4be3-9c1b-cee034e321e4 req-cd80ddeb-c73c-4a2a-a90b-37b33c28d385 service nova] Acquiring lock "refresh_cache-be825f62-cb56-4647-be1c-1605536fdc89" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.899027] env[61867]: DEBUG oslo_concurrency.lockutils [req-aaa30d77-1336-4be3-9c1b-cee034e321e4 req-cd80ddeb-c73c-4a2a-a90b-37b33c28d385 service nova] Acquired lock "refresh_cache-be825f62-cb56-4647-be1c-1605536fdc89" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.899195] env[61867]: DEBUG nova.network.neutron [req-aaa30d77-1336-4be3-9c1b-cee034e321e4 req-cd80ddeb-c73c-4a2a-a90b-37b33c28d385 service nova] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Refreshing network info cache for port d9ab90ee-f219-4de7-af2a-3db270f0f2c6 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 793.900717] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a068b31-81ee-4a3d-80a3-536601922f63 tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "905ee661-d5bd-4121-9bd0-fe68bdabd2fd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.297s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.014105] env[61867]: DEBUG nova.scheduler.client.report [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 794.202472] env[61867]: DEBUG oslo_concurrency.lockutils [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "refresh_cache-be825f62-cb56-4647-be1c-1605536fdc89" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.298120] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276473, 'name': Rename_Task, 'duration_secs': 0.129179} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.298408] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 794.298641] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-01b9e9d6-1559-409a-9d75-057997e83677 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.304888] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 794.304888] env[61867]: value = "task-1276474" [ 794.304888] env[61867]: _type = "Task" [ 794.304888] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.312473] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276474, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.404334] env[61867]: DEBUG nova.compute.manager [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 794.433617] env[61867]: DEBUG nova.network.neutron [req-aaa30d77-1336-4be3-9c1b-cee034e321e4 req-cd80ddeb-c73c-4a2a-a90b-37b33c28d385 service nova] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.518507] env[61867]: DEBUG oslo_concurrency.lockutils [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.519063] env[61867]: DEBUG nova.compute.manager [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 794.521603] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.222s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.523203] env[61867]: INFO nova.compute.claims [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 794.558446] env[61867]: DEBUG nova.network.neutron [req-aaa30d77-1336-4be3-9c1b-cee034e321e4 req-cd80ddeb-c73c-4a2a-a90b-37b33c28d385 service nova] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.644476] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1fedc835-d85a-41bb-89b3-b5da2663695e tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquiring lock "905ee661-d5bd-4121-9bd0-fe68bdabd2fd" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.644754] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1fedc835-d85a-41bb-89b3-b5da2663695e tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "905ee661-d5bd-4121-9bd0-fe68bdabd2fd" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.644997] env[61867]: INFO nova.compute.manager [None req-1fedc835-d85a-41bb-89b3-b5da2663695e tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Rebooting instance [ 794.815193] env[61867]: DEBUG oslo_vmware.api [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276474, 'name': PowerOnVM_Task, 'duration_secs': 0.417492} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.815534] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 794.815760] env[61867]: INFO nova.compute.manager [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Took 6.99 seconds to spawn the instance on the hypervisor. [ 794.815968] env[61867]: DEBUG nova.compute.manager [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 794.816826] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7926f1e0-e4a3-47c5-a3ba-931f93e75821 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.927912] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.028249] env[61867]: DEBUG nova.compute.utils [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 795.031654] env[61867]: DEBUG nova.compute.manager [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 795.031835] env[61867]: DEBUG nova.network.neutron [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 795.061546] env[61867]: DEBUG oslo_concurrency.lockutils [req-aaa30d77-1336-4be3-9c1b-cee034e321e4 req-cd80ddeb-c73c-4a2a-a90b-37b33c28d385 service nova] Releasing lock "refresh_cache-be825f62-cb56-4647-be1c-1605536fdc89" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.061966] env[61867]: DEBUG oslo_concurrency.lockutils [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "refresh_cache-be825f62-cb56-4647-be1c-1605536fdc89" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.062171] env[61867]: DEBUG nova.network.neutron [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.073076] env[61867]: DEBUG nova.policy [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc358009ebe6495a881034439d00978f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a29c5e9cfdaa4cc88ed300100a308ffa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 795.170531] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1fedc835-d85a-41bb-89b3-b5da2663695e tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquiring lock "refresh_cache-905ee661-d5bd-4121-9bd0-fe68bdabd2fd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.170643] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1fedc835-d85a-41bb-89b3-b5da2663695e tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquired lock "refresh_cache-905ee661-d5bd-4121-9bd0-fe68bdabd2fd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.170880] env[61867]: DEBUG nova.network.neutron [None req-1fedc835-d85a-41bb-89b3-b5da2663695e tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.336668] env[61867]: INFO nova.compute.manager [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Took 24.89 seconds to build instance. [ 795.530655] env[61867]: DEBUG nova.network.neutron [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Successfully created port: 6e8c2dc9-8c61-4169-928a-18c13117f16f {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 795.532970] env[61867]: DEBUG nova.compute.manager [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 795.606206] env[61867]: DEBUG nova.network.neutron [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.788241] env[61867]: DEBUG nova.network.neutron [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Updating instance_info_cache with network_info: [{"id": "d9ab90ee-f219-4de7-af2a-3db270f0f2c6", "address": "fa:16:3e:03:30:22", "network": {"id": "14cee476-296a-4093-99ba-c2a5ef63f7cf", "bridge": "br-int", "label": "tempest-ServersTestJSON-243183479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7316e4f263a9432ab2f9f91484d62f58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9ab90ee-f2", "ovs_interfaceid": "d9ab90ee-f219-4de7-af2a-3db270f0f2c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.835285] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2bb18b-f46e-4036-893b-159df3118a3c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.840140] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ad05a455-7139-4974-bc4d-ea1d8541f60a tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "3022ea79-7c27-40a4-80d6-1a86b354917a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.741s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.846454] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9cedc26-8430-4df5-b0fe-abc73784fbe3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.878599] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-173a792f-966a-4dc4-bae9-596ad60c9113 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.886432] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d55d42ae-da60-487f-aeec-b482e579e17e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.904699] env[61867]: DEBUG nova.compute.provider_tree [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.948655] env[61867]: DEBUG nova.network.neutron [None req-1fedc835-d85a-41bb-89b3-b5da2663695e tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Updating instance_info_cache with network_info: [{"id": "32f22e84-5091-47cf-a1be-91303f571008", "address": "fa:16:3e:ad:02:08", "network": {"id": "55ecbbe4-0b61-4064-8187-e30c1e444949", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-176665084-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f840df7fdbbd499cb6a10bf63ad3c837", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbdab640-5fea-4254-8bd3-f855b7eaca0d", "external-id": "nsx-vlan-transportzone-615", "segmentation_id": 615, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32f22e84-50", "ovs_interfaceid": "32f22e84-5091-47cf-a1be-91303f571008", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.987936] env[61867]: DEBUG nova.compute.manager [req-abea3c62-70c6-41d5-85dd-fc2083afc65d req-8e83c436-e609-49ea-85b0-59337b884a01 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Received event network-changed-32f22e84-5091-47cf-a1be-91303f571008 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.988171] env[61867]: DEBUG nova.compute.manager [req-abea3c62-70c6-41d5-85dd-fc2083afc65d req-8e83c436-e609-49ea-85b0-59337b884a01 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Refreshing instance network info cache due to event network-changed-32f22e84-5091-47cf-a1be-91303f571008. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 795.988649] env[61867]: DEBUG oslo_concurrency.lockutils [req-abea3c62-70c6-41d5-85dd-fc2083afc65d req-8e83c436-e609-49ea-85b0-59337b884a01 service nova] Acquiring lock "refresh_cache-905ee661-d5bd-4121-9bd0-fe68bdabd2fd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.271323] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "3022ea79-7c27-40a4-80d6-1a86b354917a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.271632] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "3022ea79-7c27-40a4-80d6-1a86b354917a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.271844] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "3022ea79-7c27-40a4-80d6-1a86b354917a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.272046] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "3022ea79-7c27-40a4-80d6-1a86b354917a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.272224] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "3022ea79-7c27-40a4-80d6-1a86b354917a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.275113] env[61867]: INFO nova.compute.manager [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Terminating instance [ 796.277873] env[61867]: DEBUG nova.compute.manager [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 796.278213] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 796.279408] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af0f3da-3a22-4d7e-b005-cc0af3f6d31e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.288746] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 796.289060] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9dfa7429-5067-436c-9356-51ff97b52978 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.291286] env[61867]: DEBUG oslo_concurrency.lockutils [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "refresh_cache-be825f62-cb56-4647-be1c-1605536fdc89" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.291637] env[61867]: DEBUG nova.compute.manager [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Instance network_info: |[{"id": "d9ab90ee-f219-4de7-af2a-3db270f0f2c6", "address": "fa:16:3e:03:30:22", "network": {"id": "14cee476-296a-4093-99ba-c2a5ef63f7cf", "bridge": "br-int", "label": "tempest-ServersTestJSON-243183479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7316e4f263a9432ab2f9f91484d62f58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9ab90ee-f2", "ovs_interfaceid": "d9ab90ee-f219-4de7-af2a-3db270f0f2c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 796.292106] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:30:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '085fb0ff-9285-4f1d-a008-a14da4844357', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd9ab90ee-f219-4de7-af2a-3db270f0f2c6', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 796.303439] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Creating folder: Project (7316e4f263a9432ab2f9f91484d62f58). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 796.304536] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0be5ae3b-1da7-4424-8a81-0b542fe53797 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.309136] env[61867]: DEBUG oslo_vmware.api [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 796.309136] env[61867]: value = "task-1276475" [ 796.309136] env[61867]: _type = "Task" [ 796.309136] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.315835] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Created folder: Project (7316e4f263a9432ab2f9f91484d62f58) in parent group-v274258. [ 796.315835] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Creating folder: Instances. Parent ref: group-v274306. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 796.319023] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9bccba11-fb27-4428-a432-7ebd253504a3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.320885] env[61867]: DEBUG oslo_vmware.api [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276475, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.328383] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Created folder: Instances in parent group-v274306. [ 796.328659] env[61867]: DEBUG oslo.service.loopingcall [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 796.328912] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 796.329203] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e82958d0-5171-420b-bc46-cceec7cab48d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.345926] env[61867]: DEBUG nova.compute.manager [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 796.350382] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 796.350382] env[61867]: value = "task-1276478" [ 796.350382] env[61867]: _type = "Task" [ 796.350382] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.357136] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276478, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.407880] env[61867]: DEBUG nova.scheduler.client.report [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 796.450903] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1fedc835-d85a-41bb-89b3-b5da2663695e tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Releasing lock "refresh_cache-905ee661-d5bd-4121-9bd0-fe68bdabd2fd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.452762] env[61867]: DEBUG oslo_concurrency.lockutils [req-abea3c62-70c6-41d5-85dd-fc2083afc65d req-8e83c436-e609-49ea-85b0-59337b884a01 service nova] Acquired lock "refresh_cache-905ee661-d5bd-4121-9bd0-fe68bdabd2fd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.453075] env[61867]: DEBUG nova.network.neutron [req-abea3c62-70c6-41d5-85dd-fc2083afc65d req-8e83c436-e609-49ea-85b0-59337b884a01 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Refreshing network info cache for port 32f22e84-5091-47cf-a1be-91303f571008 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 796.455028] env[61867]: DEBUG nova.compute.manager [None req-1fedc835-d85a-41bb-89b3-b5da2663695e tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 796.455444] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c900e319-02d5-42a6-b55a-5e002e3efbc6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.549396] env[61867]: DEBUG nova.compute.manager [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 796.574981] env[61867]: DEBUG nova.virt.hardware [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 796.575816] env[61867]: DEBUG nova.virt.hardware [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 796.575816] env[61867]: DEBUG nova.virt.hardware [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 796.575816] env[61867]: DEBUG nova.virt.hardware [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 796.575816] env[61867]: DEBUG nova.virt.hardware [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 796.576226] env[61867]: DEBUG nova.virt.hardware [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 796.576226] env[61867]: DEBUG nova.virt.hardware [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 796.576395] env[61867]: DEBUG nova.virt.hardware [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 796.576599] env[61867]: DEBUG nova.virt.hardware [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 796.576763] env[61867]: DEBUG nova.virt.hardware [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 796.576967] env[61867]: DEBUG nova.virt.hardware [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 796.577932] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610a3c5b-0555-408b-8d17-491667c40b1d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.586331] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b49f63e9-33d4-4a5b-b396-8bbda6d2c208 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.820744] env[61867]: DEBUG oslo_vmware.api [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276475, 'name': PowerOffVM_Task, 'duration_secs': 0.219647} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.821025] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 796.821202] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 796.821442] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-85066d98-3cb8-40e4-a4f6-79936c517229 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.863355] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276478, 'name': CreateVM_Task, 'duration_secs': 0.337262} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.863428] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 796.864161] env[61867]: DEBUG oslo_concurrency.lockutils [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.864335] env[61867]: DEBUG oslo_concurrency.lockutils [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.864647] env[61867]: DEBUG oslo_concurrency.lockutils [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 796.864896] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34d9bf57-4e90-43f4-85c2-e44b930ed6f4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.869482] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 796.869482] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52cae0da-1e1b-e341-3120-fc86a5854ccb" [ 796.869482] env[61867]: _type = "Task" [ 796.869482] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.873260] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.878097] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52cae0da-1e1b-e341-3120-fc86a5854ccb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.879234] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 796.879422] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 796.879595] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleting the datastore file [datastore2] 3022ea79-7c27-40a4-80d6-1a86b354917a {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 796.879892] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-89d31e72-7bef-44da-9f8f-32f3fa4d1e0f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.885675] env[61867]: DEBUG oslo_vmware.api [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 796.885675] env[61867]: value = "task-1276480" [ 796.885675] env[61867]: _type = "Task" [ 796.885675] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.892839] env[61867]: DEBUG oslo_vmware.api [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276480, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.912718] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.391s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.913195] env[61867]: DEBUG nova.compute.manager [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 796.915774] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.490s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.917120] env[61867]: INFO nova.compute.claims [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 797.151100] env[61867]: DEBUG nova.network.neutron [req-abea3c62-70c6-41d5-85dd-fc2083afc65d req-8e83c436-e609-49ea-85b0-59337b884a01 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Updated VIF entry in instance network info cache for port 32f22e84-5091-47cf-a1be-91303f571008. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 797.151457] env[61867]: DEBUG nova.network.neutron [req-abea3c62-70c6-41d5-85dd-fc2083afc65d req-8e83c436-e609-49ea-85b0-59337b884a01 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Updating instance_info_cache with network_info: [{"id": "32f22e84-5091-47cf-a1be-91303f571008", "address": "fa:16:3e:ad:02:08", "network": {"id": "55ecbbe4-0b61-4064-8187-e30c1e444949", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-176665084-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f840df7fdbbd499cb6a10bf63ad3c837", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbdab640-5fea-4254-8bd3-f855b7eaca0d", "external-id": "nsx-vlan-transportzone-615", "segmentation_id": 615, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32f22e84-50", "ovs_interfaceid": "32f22e84-5091-47cf-a1be-91303f571008", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.383549] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52cae0da-1e1b-e341-3120-fc86a5854ccb, 'name': SearchDatastore_Task, 'duration_secs': 0.010857} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.385526] env[61867]: DEBUG oslo_concurrency.lockutils [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.385526] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 797.385526] env[61867]: DEBUG oslo_concurrency.lockutils [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.385526] env[61867]: DEBUG oslo_concurrency.lockutils [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.385526] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 797.385795] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dbfddce5-e0f6-445c-b453-02f9c647a9f2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.394992] env[61867]: DEBUG oslo_vmware.api [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276480, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147363} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.396141] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 797.396391] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 797.396599] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 797.396807] env[61867]: INFO nova.compute.manager [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 797.397078] env[61867]: DEBUG oslo.service.loopingcall [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.397324] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 797.397545] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 797.398237] env[61867]: DEBUG nova.compute.manager [-] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 797.398379] env[61867]: DEBUG nova.network.neutron [-] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 797.399882] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b076642e-ab35-4830-886e-ec0966292959 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.405242] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 797.405242] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52285df0-e703-e941-6318-8340daf818cc" [ 797.405242] env[61867]: _type = "Task" [ 797.405242] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.415221] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52285df0-e703-e941-6318-8340daf818cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.427139] env[61867]: DEBUG nova.compute.utils [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 797.427139] env[61867]: DEBUG nova.compute.manager [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 797.427139] env[61867]: DEBUG nova.network.neutron [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 797.471083] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa3d037-b83d-4d5e-ae8d-2fad28215ddd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.478388] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1fedc835-d85a-41bb-89b3-b5da2663695e tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Doing hard reboot of VM {{(pid=61867) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 797.478633] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-bedee612-c609-4850-98de-9744cc0ca33f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.487120] env[61867]: DEBUG nova.network.neutron [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Successfully updated port: 6e8c2dc9-8c61-4169-928a-18c13117f16f {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 797.491701] env[61867]: DEBUG oslo_vmware.api [None req-1fedc835-d85a-41bb-89b3-b5da2663695e tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Waiting for the task: (returnval){ [ 797.491701] env[61867]: value = "task-1276481" [ 797.491701] env[61867]: _type = "Task" [ 797.491701] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.491701] env[61867]: DEBUG nova.policy [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32f3a7fca4774948a71bdded455edcde', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c8e9d6e6ceac4520add365429c4eb4d2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 797.503963] env[61867]: DEBUG oslo_vmware.api [None req-1fedc835-d85a-41bb-89b3-b5da2663695e tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276481, 'name': ResetVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.655689] env[61867]: DEBUG oslo_concurrency.lockutils [req-abea3c62-70c6-41d5-85dd-fc2083afc65d req-8e83c436-e609-49ea-85b0-59337b884a01 service nova] Releasing lock "refresh_cache-905ee661-d5bd-4121-9bd0-fe68bdabd2fd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.813320] env[61867]: DEBUG nova.compute.manager [req-75ac5029-4545-4e8e-bdcf-11d0794723de req-8aa9031e-e2f0-477e-a007-68154dadadd6 service nova] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Received event network-vif-deleted-8546f778-9a66-4428-a784-38b85f0f3f98 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 797.815393] env[61867]: INFO nova.compute.manager [req-75ac5029-4545-4e8e-bdcf-11d0794723de req-8aa9031e-e2f0-477e-a007-68154dadadd6 service nova] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Neutron deleted interface 8546f778-9a66-4428-a784-38b85f0f3f98; detaching it from the instance and deleting it from the info cache [ 797.815603] env[61867]: DEBUG nova.network.neutron [req-75ac5029-4545-4e8e-bdcf-11d0794723de req-8aa9031e-e2f0-477e-a007-68154dadadd6 service nova] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.898325] env[61867]: DEBUG nova.network.neutron [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Successfully created port: 943c04a5-7eba-46db-ab4f-93f7f84be14e {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 797.914630] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52285df0-e703-e941-6318-8340daf818cc, 'name': SearchDatastore_Task, 'duration_secs': 0.008301} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.915433] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b23afdbc-5fb8-4245-8ebf-a8b29bc76cb5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.920643] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 797.920643] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527c928c-0b75-310a-45f4-6faf1b4d6374" [ 797.920643] env[61867]: _type = "Task" [ 797.920643] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.930789] env[61867]: DEBUG nova.compute.manager [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 797.933378] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527c928c-0b75-310a-45f4-6faf1b4d6374, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.991456] env[61867]: DEBUG oslo_concurrency.lockutils [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "refresh_cache-80a703b3-d692-4023-a73b-dba980a94dff" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.991637] env[61867]: DEBUG oslo_concurrency.lockutils [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquired lock "refresh_cache-80a703b3-d692-4023-a73b-dba980a94dff" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.991802] env[61867]: DEBUG nova.network.neutron [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 798.008568] env[61867]: DEBUG oslo_vmware.api [None req-1fedc835-d85a-41bb-89b3-b5da2663695e tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276481, 'name': ResetVM_Task, 'duration_secs': 0.165611} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.008840] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1fedc835-d85a-41bb-89b3-b5da2663695e tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Did hard reboot of VM {{(pid=61867) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 798.009075] env[61867]: DEBUG nova.compute.manager [None req-1fedc835-d85a-41bb-89b3-b5da2663695e tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 798.009814] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a307134e-f354-4164-9f4d-f8cb008a9e54 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.026817] env[61867]: DEBUG nova.compute.manager [req-1f5d5437-3b1f-4717-878d-4ff899a2e9d1 req-7bb8036c-a7df-467a-b69a-ceafa1c68bd2 service nova] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Received event network-vif-plugged-6e8c2dc9-8c61-4169-928a-18c13117f16f {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 798.026817] env[61867]: DEBUG oslo_concurrency.lockutils [req-1f5d5437-3b1f-4717-878d-4ff899a2e9d1 req-7bb8036c-a7df-467a-b69a-ceafa1c68bd2 service nova] Acquiring lock "80a703b3-d692-4023-a73b-dba980a94dff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.026990] env[61867]: DEBUG oslo_concurrency.lockutils [req-1f5d5437-3b1f-4717-878d-4ff899a2e9d1 req-7bb8036c-a7df-467a-b69a-ceafa1c68bd2 service nova] Lock "80a703b3-d692-4023-a73b-dba980a94dff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.027174] env[61867]: DEBUG oslo_concurrency.lockutils [req-1f5d5437-3b1f-4717-878d-4ff899a2e9d1 req-7bb8036c-a7df-467a-b69a-ceafa1c68bd2 service nova] Lock "80a703b3-d692-4023-a73b-dba980a94dff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.027290] env[61867]: DEBUG nova.compute.manager [req-1f5d5437-3b1f-4717-878d-4ff899a2e9d1 req-7bb8036c-a7df-467a-b69a-ceafa1c68bd2 service nova] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] No waiting events found dispatching network-vif-plugged-6e8c2dc9-8c61-4169-928a-18c13117f16f {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 798.027506] env[61867]: WARNING nova.compute.manager [req-1f5d5437-3b1f-4717-878d-4ff899a2e9d1 req-7bb8036c-a7df-467a-b69a-ceafa1c68bd2 service nova] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Received unexpected event network-vif-plugged-6e8c2dc9-8c61-4169-928a-18c13117f16f for instance with vm_state building and task_state spawning. [ 798.027604] env[61867]: DEBUG nova.compute.manager [req-1f5d5437-3b1f-4717-878d-4ff899a2e9d1 req-7bb8036c-a7df-467a-b69a-ceafa1c68bd2 service nova] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Received event network-changed-6e8c2dc9-8c61-4169-928a-18c13117f16f {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 798.027763] env[61867]: DEBUG nova.compute.manager [req-1f5d5437-3b1f-4717-878d-4ff899a2e9d1 req-7bb8036c-a7df-467a-b69a-ceafa1c68bd2 service nova] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Refreshing instance network info cache due to event network-changed-6e8c2dc9-8c61-4169-928a-18c13117f16f. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 798.027907] env[61867]: DEBUG oslo_concurrency.lockutils [req-1f5d5437-3b1f-4717-878d-4ff899a2e9d1 req-7bb8036c-a7df-467a-b69a-ceafa1c68bd2 service nova] Acquiring lock "refresh_cache-80a703b3-d692-4023-a73b-dba980a94dff" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.171836] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715ce5df-2f99-4413-8c8f-b47f11d27cc0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.179829] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77be1292-75f9-4258-b684-737a097d9051 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.211050] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b729fc8-4444-4ee2-a047-e2c451352fb8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.218664] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b1f67b-ff55-48d5-8493-e69cca1d8241 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.232476] env[61867]: DEBUG nova.compute.provider_tree [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.287036] env[61867]: DEBUG nova.network.neutron [-] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.318617] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-81e1330e-5072-4797-83c5-ddc919488ee9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.328062] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f35399-e7a0-4dbb-b33c-114470f4ef17 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.349910] env[61867]: DEBUG nova.compute.manager [req-75ac5029-4545-4e8e-bdcf-11d0794723de req-8aa9031e-e2f0-477e-a007-68154dadadd6 service nova] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Detach interface failed, port_id=8546f778-9a66-4428-a784-38b85f0f3f98, reason: Instance 3022ea79-7c27-40a4-80d6-1a86b354917a could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 798.432676] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527c928c-0b75-310a-45f4-6faf1b4d6374, 'name': SearchDatastore_Task, 'duration_secs': 0.008652} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.433217] env[61867]: DEBUG oslo_concurrency.lockutils [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.433531] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] be825f62-cb56-4647-be1c-1605536fdc89/be825f62-cb56-4647-be1c-1605536fdc89.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 798.433809] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d3e054ed-befb-4a83-a766-cd82bbdbf535 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.444505] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 798.444505] env[61867]: value = "task-1276482" [ 798.444505] env[61867]: _type = "Task" [ 798.444505] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.452891] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276482, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.526282] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1fedc835-d85a-41bb-89b3-b5da2663695e tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "905ee661-d5bd-4121-9bd0-fe68bdabd2fd" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.881s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.533978] env[61867]: DEBUG nova.network.neutron [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 798.678716] env[61867]: DEBUG nova.network.neutron [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Updating instance_info_cache with network_info: [{"id": "6e8c2dc9-8c61-4169-928a-18c13117f16f", "address": "fa:16:3e:8c:a3:15", "network": {"id": "a645da41-1400-470e-acf2-68cec86432fd", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1387586368-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a29c5e9cfdaa4cc88ed300100a308ffa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e8c2dc9-8c", "ovs_interfaceid": "6e8c2dc9-8c61-4169-928a-18c13117f16f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.735566] env[61867]: DEBUG nova.scheduler.client.report [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 798.790160] env[61867]: INFO nova.compute.manager [-] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Took 1.39 seconds to deallocate network for instance. [ 798.941483] env[61867]: DEBUG nova.compute.manager [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 798.956706] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276482, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451554} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.956706] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] be825f62-cb56-4647-be1c-1605536fdc89/be825f62-cb56-4647-be1c-1605536fdc89.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 798.956706] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 798.956944] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-04901a89-eeec-40f7-a25f-3a6a9c5f9c8c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.965242] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 798.965242] env[61867]: value = "task-1276483" [ 798.965242] env[61867]: _type = "Task" [ 798.965242] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.967284] env[61867]: DEBUG nova.virt.hardware [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 798.967939] env[61867]: DEBUG nova.virt.hardware [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 798.967939] env[61867]: DEBUG nova.virt.hardware [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.967939] env[61867]: DEBUG nova.virt.hardware [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 798.968112] env[61867]: DEBUG nova.virt.hardware [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.968178] env[61867]: DEBUG nova.virt.hardware [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 798.968381] env[61867]: DEBUG nova.virt.hardware [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 798.968541] env[61867]: DEBUG nova.virt.hardware [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 798.968704] env[61867]: DEBUG nova.virt.hardware [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 798.968894] env[61867]: DEBUG nova.virt.hardware [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 798.969198] env[61867]: DEBUG nova.virt.hardware [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 798.969948] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b15dcfe-530a-4583-9a21-c29182940791 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.985301] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44259283-ca85-4ca3-bf70-3564f873f52d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.989391] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276483, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.181547] env[61867]: DEBUG oslo_concurrency.lockutils [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Releasing lock "refresh_cache-80a703b3-d692-4023-a73b-dba980a94dff" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.181899] env[61867]: DEBUG nova.compute.manager [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Instance network_info: |[{"id": "6e8c2dc9-8c61-4169-928a-18c13117f16f", "address": "fa:16:3e:8c:a3:15", "network": {"id": "a645da41-1400-470e-acf2-68cec86432fd", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1387586368-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a29c5e9cfdaa4cc88ed300100a308ffa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e8c2dc9-8c", "ovs_interfaceid": "6e8c2dc9-8c61-4169-928a-18c13117f16f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 799.182219] env[61867]: DEBUG oslo_concurrency.lockutils [req-1f5d5437-3b1f-4717-878d-4ff899a2e9d1 req-7bb8036c-a7df-467a-b69a-ceafa1c68bd2 service nova] Acquired lock "refresh_cache-80a703b3-d692-4023-a73b-dba980a94dff" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.182431] env[61867]: DEBUG nova.network.neutron [req-1f5d5437-3b1f-4717-878d-4ff899a2e9d1 req-7bb8036c-a7df-467a-b69a-ceafa1c68bd2 service nova] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Refreshing network info cache for port 6e8c2dc9-8c61-4169-928a-18c13117f16f {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 799.183636] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8c:a3:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d377d75-3add-4a15-8691-74b2eb010924', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6e8c2dc9-8c61-4169-928a-18c13117f16f', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 799.191707] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Creating folder: Project (a29c5e9cfdaa4cc88ed300100a308ffa). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 799.194753] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-929aa4e1-54d8-434e-b4f2-e2f8ffa79891 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.205765] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Created folder: Project (a29c5e9cfdaa4cc88ed300100a308ffa) in parent group-v274258. [ 799.206063] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Creating folder: Instances. Parent ref: group-v274309. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 799.206223] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b448607-7844-43c7-b389-03b26f40f67d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.214908] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Created folder: Instances in parent group-v274309. [ 799.215321] env[61867]: DEBUG oslo.service.loopingcall [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 799.215321] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 799.215549] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7d4acda8-f7d6-46e9-9ea1-cd9f3c22e2c5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.236084] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 799.236084] env[61867]: value = "task-1276486" [ 799.236084] env[61867]: _type = "Task" [ 799.236084] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.240246] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.324s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.240795] env[61867]: DEBUG nova.compute.manager [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 799.252281] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.176s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.252548] env[61867]: DEBUG nova.objects.instance [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lazy-loading 'resources' on Instance uuid b5a6ff7c-2b48-4f82-ba53-ec1977736c8f {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 799.253727] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276486, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.302019] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.479593] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276483, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058552} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.479879] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 799.481984] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1315e83c-8379-442d-9700-6ee7e4b090c3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.502597] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] be825f62-cb56-4647-be1c-1605536fdc89/be825f62-cb56-4647-be1c-1605536fdc89.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 799.503715] env[61867]: DEBUG nova.network.neutron [req-1f5d5437-3b1f-4717-878d-4ff899a2e9d1 req-7bb8036c-a7df-467a-b69a-ceafa1c68bd2 service nova] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Updated VIF entry in instance network info cache for port 6e8c2dc9-8c61-4169-928a-18c13117f16f. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 799.504135] env[61867]: DEBUG nova.network.neutron [req-1f5d5437-3b1f-4717-878d-4ff899a2e9d1 req-7bb8036c-a7df-467a-b69a-ceafa1c68bd2 service nova] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Updating instance_info_cache with network_info: [{"id": "6e8c2dc9-8c61-4169-928a-18c13117f16f", "address": "fa:16:3e:8c:a3:15", "network": {"id": "a645da41-1400-470e-acf2-68cec86432fd", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1387586368-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a29c5e9cfdaa4cc88ed300100a308ffa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e8c2dc9-8c", "ovs_interfaceid": "6e8c2dc9-8c61-4169-928a-18c13117f16f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.506150] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cca856f6-f2f7-42ff-bdce-f7dffd113200 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.521864] env[61867]: DEBUG nova.compute.manager [req-46a22237-053e-4def-8283-93a021b724f6 req-a8a9c5ec-c0f0-4b39-87d0-f00b13843b18 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Received event network-vif-plugged-943c04a5-7eba-46db-ab4f-93f7f84be14e {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.522136] env[61867]: DEBUG oslo_concurrency.lockutils [req-46a22237-053e-4def-8283-93a021b724f6 req-a8a9c5ec-c0f0-4b39-87d0-f00b13843b18 service nova] Acquiring lock "1df8427c-e75d-4b60-a92a-b5ba76b67081-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.522356] env[61867]: DEBUG oslo_concurrency.lockutils [req-46a22237-053e-4def-8283-93a021b724f6 req-a8a9c5ec-c0f0-4b39-87d0-f00b13843b18 service nova] Lock "1df8427c-e75d-4b60-a92a-b5ba76b67081-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.522520] env[61867]: DEBUG oslo_concurrency.lockutils [req-46a22237-053e-4def-8283-93a021b724f6 req-a8a9c5ec-c0f0-4b39-87d0-f00b13843b18 service nova] Lock "1df8427c-e75d-4b60-a92a-b5ba76b67081-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.522686] env[61867]: DEBUG nova.compute.manager [req-46a22237-053e-4def-8283-93a021b724f6 req-a8a9c5ec-c0f0-4b39-87d0-f00b13843b18 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] No waiting events found dispatching network-vif-plugged-943c04a5-7eba-46db-ab4f-93f7f84be14e {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 799.522846] env[61867]: WARNING nova.compute.manager [req-46a22237-053e-4def-8283-93a021b724f6 req-a8a9c5ec-c0f0-4b39-87d0-f00b13843b18 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Received unexpected event network-vif-plugged-943c04a5-7eba-46db-ab4f-93f7f84be14e for instance with vm_state building and task_state spawning. [ 799.529363] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 799.529363] env[61867]: value = "task-1276487" [ 799.529363] env[61867]: _type = "Task" [ 799.529363] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.537572] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276487, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.746206] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276486, 'name': CreateVM_Task, 'duration_secs': 0.328239} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.746262] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 799.747036] env[61867]: DEBUG oslo_concurrency.lockutils [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.747210] env[61867]: DEBUG oslo_concurrency.lockutils [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.747561] env[61867]: DEBUG oslo_concurrency.lockutils [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 799.747811] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d97ae0b5-32b8-4bfd-a80d-dd9e3726a22a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.753454] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 799.753454] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526e1211-3057-0051-6970-c7896162d664" [ 799.753454] env[61867]: _type = "Task" [ 799.753454] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.758528] env[61867]: DEBUG nova.compute.utils [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 799.763071] env[61867]: DEBUG nova.compute.manager [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 799.763260] env[61867]: DEBUG nova.network.neutron [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 799.775772] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526e1211-3057-0051-6970-c7896162d664, 'name': SearchDatastore_Task, 'duration_secs': 0.009846} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.775772] env[61867]: DEBUG oslo_concurrency.lockutils [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.775772] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.775772] env[61867]: DEBUG oslo_concurrency.lockutils [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.775924] env[61867]: DEBUG oslo_concurrency.lockutils [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.775924] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.775924] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4194c3fa-3159-410e-b26b-591e3e8ea961 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.782647] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.782849] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 799.784239] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29142fb8-cbfd-4837-ba1f-f2589375ab1a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.792147] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 799.792147] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5253e8bb-c3b9-b86c-52e4-b2b769e0762f" [ 799.792147] env[61867]: _type = "Task" [ 799.792147] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.799945] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5253e8bb-c3b9-b86c-52e4-b2b769e0762f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.819469] env[61867]: DEBUG nova.policy [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e18b96b07e54a1d8ed642b02f11e3a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b64ff57036a492eb2555f33e0af664e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 799.983878] env[61867]: DEBUG nova.network.neutron [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Successfully updated port: 943c04a5-7eba-46db-ab4f-93f7f84be14e {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 800.013383] env[61867]: DEBUG nova.compute.manager [req-88efe97e-fd1d-4fae-a7ca-57c217eb2f81 req-fb244a3b-881b-441c-96d0-6cbd367680f5 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Received event network-changed-943c04a5-7eba-46db-ab4f-93f7f84be14e {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 800.013585] env[61867]: DEBUG nova.compute.manager [req-88efe97e-fd1d-4fae-a7ca-57c217eb2f81 req-fb244a3b-881b-441c-96d0-6cbd367680f5 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Refreshing instance network info cache due to event network-changed-943c04a5-7eba-46db-ab4f-93f7f84be14e. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 800.013796] env[61867]: DEBUG oslo_concurrency.lockutils [req-88efe97e-fd1d-4fae-a7ca-57c217eb2f81 req-fb244a3b-881b-441c-96d0-6cbd367680f5 service nova] Acquiring lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.013949] env[61867]: DEBUG oslo_concurrency.lockutils [req-88efe97e-fd1d-4fae-a7ca-57c217eb2f81 req-fb244a3b-881b-441c-96d0-6cbd367680f5 service nova] Acquired lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.014248] env[61867]: DEBUG nova.network.neutron [req-88efe97e-fd1d-4fae-a7ca-57c217eb2f81 req-fb244a3b-881b-441c-96d0-6cbd367680f5 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Refreshing network info cache for port 943c04a5-7eba-46db-ab4f-93f7f84be14e {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 800.023843] env[61867]: DEBUG oslo_concurrency.lockutils [req-1f5d5437-3b1f-4717-878d-4ff899a2e9d1 req-7bb8036c-a7df-467a-b69a-ceafa1c68bd2 service nova] Releasing lock "refresh_cache-80a703b3-d692-4023-a73b-dba980a94dff" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.041546] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276487, 'name': ReconfigVM_Task, 'duration_secs': 0.264209} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.042297] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Reconfigured VM instance instance-0000003c to attach disk [datastore2] be825f62-cb56-4647-be1c-1605536fdc89/be825f62-cb56-4647-be1c-1605536fdc89.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 800.042457] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dffa1a4a-f6bf-453c-bacd-d7bca58731ef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.048280] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cfb4232-4afa-4025-8bf3-66fd31a2d34a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.052922] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 800.052922] env[61867]: value = "task-1276488" [ 800.052922] env[61867]: _type = "Task" [ 800.052922] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.058893] env[61867]: DEBUG oslo_concurrency.lockutils [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquiring lock "905ee661-d5bd-4121-9bd0-fe68bdabd2fd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.059148] env[61867]: DEBUG oslo_concurrency.lockutils [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "905ee661-d5bd-4121-9bd0-fe68bdabd2fd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.059355] env[61867]: DEBUG oslo_concurrency.lockutils [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquiring lock "905ee661-d5bd-4121-9bd0-fe68bdabd2fd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.059536] env[61867]: DEBUG oslo_concurrency.lockutils [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "905ee661-d5bd-4121-9bd0-fe68bdabd2fd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.059835] env[61867]: DEBUG oslo_concurrency.lockutils [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "905ee661-d5bd-4121-9bd0-fe68bdabd2fd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.061799] env[61867]: INFO nova.compute.manager [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Terminating instance [ 800.063629] env[61867]: DEBUG nova.compute.manager [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 800.063751] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 800.065716] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf084a56-43ad-4239-977e-1fd94565a118 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.069812] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b21ba97-9cfd-485b-9362-eaac7d6b8d10 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.073995] env[61867]: DEBUG nova.compute.manager [req-c9c33c0b-2c2d-490e-b37b-818cd64adb6a req-8ceb71f1-71da-43f8-8542-2fbe6239e1f8 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Received event network-changed-32f22e84-5091-47cf-a1be-91303f571008 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 800.074198] env[61867]: DEBUG nova.compute.manager [req-c9c33c0b-2c2d-490e-b37b-818cd64adb6a req-8ceb71f1-71da-43f8-8542-2fbe6239e1f8 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Refreshing instance network info cache due to event network-changed-32f22e84-5091-47cf-a1be-91303f571008. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 800.074407] env[61867]: DEBUG oslo_concurrency.lockutils [req-c9c33c0b-2c2d-490e-b37b-818cd64adb6a req-8ceb71f1-71da-43f8-8542-2fbe6239e1f8 service nova] Acquiring lock "refresh_cache-905ee661-d5bd-4121-9bd0-fe68bdabd2fd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.074906] env[61867]: DEBUG oslo_concurrency.lockutils [req-c9c33c0b-2c2d-490e-b37b-818cd64adb6a req-8ceb71f1-71da-43f8-8542-2fbe6239e1f8 service nova] Acquired lock "refresh_cache-905ee661-d5bd-4121-9bd0-fe68bdabd2fd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.075119] env[61867]: DEBUG nova.network.neutron [req-c9c33c0b-2c2d-490e-b37b-818cd64adb6a req-8ceb71f1-71da-43f8-8542-2fbe6239e1f8 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Refreshing network info cache for port 32f22e84-5091-47cf-a1be-91303f571008 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 800.080390] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276488, 'name': Rename_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.085145] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 800.114855] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5098e14d-d4c0-4d5d-8b33-57852c0ee235 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.118611] env[61867]: DEBUG nova.network.neutron [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Successfully created port: 79dddbac-ec4f-4cd9-8436-a3600957e598 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 800.120859] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d47d6e6-659e-4abe-923f-3c61e96a2471 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.130469] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f5e70e-85c1-4b13-a4c8-82db4645d26d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.134308] env[61867]: DEBUG oslo_vmware.api [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Waiting for the task: (returnval){ [ 800.134308] env[61867]: value = "task-1276489" [ 800.134308] env[61867]: _type = "Task" [ 800.134308] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.146395] env[61867]: DEBUG nova.compute.provider_tree [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.152982] env[61867]: DEBUG oslo_vmware.api [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276489, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.263506] env[61867]: DEBUG nova.compute.manager [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 800.302745] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5253e8bb-c3b9-b86c-52e4-b2b769e0762f, 'name': SearchDatastore_Task, 'duration_secs': 0.008712} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.303570] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d96a7a6-8c75-45fb-918d-88df5a421fab {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.310672] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 800.310672] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527f1993-6ba7-6e34-c707-22fd6d71a343" [ 800.310672] env[61867]: _type = "Task" [ 800.310672] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.318805] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527f1993-6ba7-6e34-c707-22fd6d71a343, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.489638] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.544076] env[61867]: DEBUG nova.network.neutron [req-88efe97e-fd1d-4fae-a7ca-57c217eb2f81 req-fb244a3b-881b-441c-96d0-6cbd367680f5 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.562803] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276488, 'name': Rename_Task, 'duration_secs': 0.171962} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.562962] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 800.563855] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-23ca3e8e-77de-4502-9f9a-eb9753c304a1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.569641] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 800.569641] env[61867]: value = "task-1276490" [ 800.569641] env[61867]: _type = "Task" [ 800.569641] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.578908] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276490, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.623048] env[61867]: DEBUG nova.network.neutron [req-88efe97e-fd1d-4fae-a7ca-57c217eb2f81 req-fb244a3b-881b-441c-96d0-6cbd367680f5 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.644919] env[61867]: DEBUG oslo_vmware.api [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276489, 'name': PowerOffVM_Task, 'duration_secs': 0.201223} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.647522] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 800.647522] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 800.647522] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-56147383-c70c-4d68-8ddf-0769610cd0bc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.652028] env[61867]: DEBUG nova.scheduler.client.report [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 800.714022] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 800.714022] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 800.714022] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Deleting the datastore file [datastore2] 905ee661-d5bd-4121-9bd0-fe68bdabd2fd {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 800.714022] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a26cb869-b447-4aa4-ba2d-f03331e69662 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.720283] env[61867]: DEBUG oslo_vmware.api [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Waiting for the task: (returnval){ [ 800.720283] env[61867]: value = "task-1276492" [ 800.720283] env[61867]: _type = "Task" [ 800.720283] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.728520] env[61867]: DEBUG oslo_vmware.api [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276492, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.822374] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527f1993-6ba7-6e34-c707-22fd6d71a343, 'name': SearchDatastore_Task, 'duration_secs': 0.009393} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.822692] env[61867]: DEBUG oslo_concurrency.lockutils [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.823040] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 80a703b3-d692-4023-a73b-dba980a94dff/80a703b3-d692-4023-a73b-dba980a94dff.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 800.823358] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-23f453a6-11b7-451f-98cd-d62eae704f43 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.830591] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 800.830591] env[61867]: value = "task-1276493" [ 800.830591] env[61867]: _type = "Task" [ 800.830591] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.840782] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276493, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.080776] env[61867]: DEBUG oslo_vmware.api [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276490, 'name': PowerOnVM_Task, 'duration_secs': 0.471055} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.081109] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 801.081370] env[61867]: INFO nova.compute.manager [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Took 8.80 seconds to spawn the instance on the hypervisor. [ 801.081550] env[61867]: DEBUG nova.compute.manager [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 801.082501] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f937d75-923a-4cf0-ab9f-1bab4d2067e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.125913] env[61867]: DEBUG oslo_concurrency.lockutils [req-88efe97e-fd1d-4fae-a7ca-57c217eb2f81 req-fb244a3b-881b-441c-96d0-6cbd367680f5 service nova] Releasing lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.126393] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquired lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.126551] env[61867]: DEBUG nova.network.neutron [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 801.156353] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.904s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.158974] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.915s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.159304] env[61867]: DEBUG nova.objects.instance [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lazy-loading 'resources' on Instance uuid 81ada1e9-b100-44fd-8119-7ef8695e601f {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 801.185600] env[61867]: INFO nova.scheduler.client.report [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Deleted allocations for instance b5a6ff7c-2b48-4f82-ba53-ec1977736c8f [ 801.208380] env[61867]: DEBUG nova.network.neutron [req-c9c33c0b-2c2d-490e-b37b-818cd64adb6a req-8ceb71f1-71da-43f8-8542-2fbe6239e1f8 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Updated VIF entry in instance network info cache for port 32f22e84-5091-47cf-a1be-91303f571008. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 801.208747] env[61867]: DEBUG nova.network.neutron [req-c9c33c0b-2c2d-490e-b37b-818cd64adb6a req-8ceb71f1-71da-43f8-8542-2fbe6239e1f8 service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Updating instance_info_cache with network_info: [{"id": "32f22e84-5091-47cf-a1be-91303f571008", "address": "fa:16:3e:ad:02:08", "network": {"id": "55ecbbe4-0b61-4064-8187-e30c1e444949", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-176665084-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f840df7fdbbd499cb6a10bf63ad3c837", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbdab640-5fea-4254-8bd3-f855b7eaca0d", "external-id": "nsx-vlan-transportzone-615", "segmentation_id": 615, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32f22e84-50", "ovs_interfaceid": "32f22e84-5091-47cf-a1be-91303f571008", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.234654] env[61867]: DEBUG oslo_vmware.api [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Task: {'id': task-1276492, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148046} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.235023] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 801.235214] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 801.235394] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 801.235568] env[61867]: INFO nova.compute.manager [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Took 1.17 seconds to destroy the instance on the hypervisor. [ 801.235891] env[61867]: DEBUG oslo.service.loopingcall [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 801.236118] env[61867]: DEBUG nova.compute.manager [-] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 801.236261] env[61867]: DEBUG nova.network.neutron [-] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 801.274159] env[61867]: DEBUG nova.compute.manager [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 801.307922] env[61867]: DEBUG nova.virt.hardware [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:47:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='f8967eee-6d82-41b2-98fd-bc50fb8c603b',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-993246556',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 801.308340] env[61867]: DEBUG nova.virt.hardware [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 801.308433] env[61867]: DEBUG nova.virt.hardware [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 801.308669] env[61867]: DEBUG nova.virt.hardware [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 801.308852] env[61867]: DEBUG nova.virt.hardware [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 801.310224] env[61867]: DEBUG nova.virt.hardware [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 801.310517] env[61867]: DEBUG nova.virt.hardware [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 801.310741] env[61867]: DEBUG nova.virt.hardware [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 801.310970] env[61867]: DEBUG nova.virt.hardware [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 801.311206] env[61867]: DEBUG nova.virt.hardware [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 801.311405] env[61867]: DEBUG nova.virt.hardware [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.312280] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2517d01-d120-4af2-958a-cd8365daf90a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.326680] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63cbce35-72f8-4eb0-8276-292c0d45dd85 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.352098] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276493, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492867} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.352637] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 80a703b3-d692-4023-a73b-dba980a94dff/80a703b3-d692-4023-a73b-dba980a94dff.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 801.352900] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 801.353416] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5da3c178-48df-471d-ac62-0cfe5bdca215 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.359752] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 801.359752] env[61867]: value = "task-1276494" [ 801.359752] env[61867]: _type = "Task" [ 801.359752] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.367819] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276494, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.606255] env[61867]: INFO nova.compute.manager [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Took 29.66 seconds to build instance. [ 801.644510] env[61867]: DEBUG nova.network.neutron [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Successfully updated port: 79dddbac-ec4f-4cd9-8436-a3600957e598 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 801.676384] env[61867]: DEBUG nova.network.neutron [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 801.693038] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e3a7d5ed-f2e0-472e-835d-9a4e701a711c tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "b5a6ff7c-2b48-4f82-ba53-ec1977736c8f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.524s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.714407] env[61867]: DEBUG oslo_concurrency.lockutils [req-c9c33c0b-2c2d-490e-b37b-818cd64adb6a req-8ceb71f1-71da-43f8-8542-2fbe6239e1f8 service nova] Releasing lock "refresh_cache-905ee661-d5bd-4121-9bd0-fe68bdabd2fd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.874964] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276494, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.4974} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.877261] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 801.878169] env[61867]: DEBUG nova.network.neutron [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Updating instance_info_cache with network_info: [{"id": "943c04a5-7eba-46db-ab4f-93f7f84be14e", "address": "fa:16:3e:56:de:39", "network": {"id": "7635ca79-1de7-4ba3-b347-1e0b66d7eb56", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1574772814-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8e9d6e6ceac4520add365429c4eb4d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d0c6fd7-3cc9-4818-9475-8f15900394cc", "external-id": "nsx-vlan-transportzone-317", "segmentation_id": 317, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap943c04a5-7e", "ovs_interfaceid": "943c04a5-7eba-46db-ab4f-93f7f84be14e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.879782] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7edc4eb-7185-4a6f-9319-5bd13760fab1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.906989] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] 80a703b3-d692-4023-a73b-dba980a94dff/80a703b3-d692-4023-a73b-dba980a94dff.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.910110] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec316040-991e-4112-9409-f326aeb814d2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.930475] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 801.930475] env[61867]: value = "task-1276495" [ 801.930475] env[61867]: _type = "Task" [ 801.930475] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.939863] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276495, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.949931] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19377dad-c88a-4c33-b108-feadf7e603e5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.956853] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e4577ba-4b79-43a1-a667-238a9411c3be {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.985667] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aeeff78-b413-4af6-93f4-9f0196a39884 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.992800] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7c5d715-d263-41c4-b8ea-47c2fb1480a0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.008574] env[61867]: DEBUG nova.compute.provider_tree [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.062630] env[61867]: DEBUG nova.compute.manager [req-2b07905b-e720-4522-95a0-2f06f9e0f10f req-e2d8fef6-85fc-4c3c-9cee-117546cdab74 service nova] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Received event network-vif-plugged-79dddbac-ec4f-4cd9-8436-a3600957e598 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.062890] env[61867]: DEBUG oslo_concurrency.lockutils [req-2b07905b-e720-4522-95a0-2f06f9e0f10f req-e2d8fef6-85fc-4c3c-9cee-117546cdab74 service nova] Acquiring lock "81e70c36-04ea-450c-9383-53ef069d1c46-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.063054] env[61867]: DEBUG oslo_concurrency.lockutils [req-2b07905b-e720-4522-95a0-2f06f9e0f10f req-e2d8fef6-85fc-4c3c-9cee-117546cdab74 service nova] Lock "81e70c36-04ea-450c-9383-53ef069d1c46-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.063217] env[61867]: DEBUG oslo_concurrency.lockutils [req-2b07905b-e720-4522-95a0-2f06f9e0f10f req-e2d8fef6-85fc-4c3c-9cee-117546cdab74 service nova] Lock "81e70c36-04ea-450c-9383-53ef069d1c46-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.063384] env[61867]: DEBUG nova.compute.manager [req-2b07905b-e720-4522-95a0-2f06f9e0f10f req-e2d8fef6-85fc-4c3c-9cee-117546cdab74 service nova] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] No waiting events found dispatching network-vif-plugged-79dddbac-ec4f-4cd9-8436-a3600957e598 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 802.063545] env[61867]: WARNING nova.compute.manager [req-2b07905b-e720-4522-95a0-2f06f9e0f10f req-e2d8fef6-85fc-4c3c-9cee-117546cdab74 service nova] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Received unexpected event network-vif-plugged-79dddbac-ec4f-4cd9-8436-a3600957e598 for instance with vm_state building and task_state spawning. [ 802.063702] env[61867]: DEBUG nova.compute.manager [req-2b07905b-e720-4522-95a0-2f06f9e0f10f req-e2d8fef6-85fc-4c3c-9cee-117546cdab74 service nova] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Received event network-changed-79dddbac-ec4f-4cd9-8436-a3600957e598 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.063853] env[61867]: DEBUG nova.compute.manager [req-2b07905b-e720-4522-95a0-2f06f9e0f10f req-e2d8fef6-85fc-4c3c-9cee-117546cdab74 service nova] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Refreshing instance network info cache due to event network-changed-79dddbac-ec4f-4cd9-8436-a3600957e598. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 802.064039] env[61867]: DEBUG oslo_concurrency.lockutils [req-2b07905b-e720-4522-95a0-2f06f9e0f10f req-e2d8fef6-85fc-4c3c-9cee-117546cdab74 service nova] Acquiring lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.064183] env[61867]: DEBUG oslo_concurrency.lockutils [req-2b07905b-e720-4522-95a0-2f06f9e0f10f req-e2d8fef6-85fc-4c3c-9cee-117546cdab74 service nova] Acquired lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.064328] env[61867]: DEBUG nova.network.neutron [req-2b07905b-e720-4522-95a0-2f06f9e0f10f req-e2d8fef6-85fc-4c3c-9cee-117546cdab74 service nova] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Refreshing network info cache for port 79dddbac-ec4f-4cd9-8436-a3600957e598 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 802.095405] env[61867]: DEBUG nova.network.neutron [-] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.105642] env[61867]: DEBUG oslo_concurrency.lockutils [None req-16132fdb-31dd-4d8b-a798-52e22e3bab18 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "be825f62-cb56-4647-be1c-1605536fdc89" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.345s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.128233] env[61867]: DEBUG nova.compute.manager [req-f200375d-7357-4f35-b0b3-1dab7ad39312 req-e7dd3084-bbb9-4054-a032-87705158243d service nova] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Received event network-vif-deleted-32f22e84-5091-47cf-a1be-91303f571008 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.147392] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.388255] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Releasing lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.388595] env[61867]: DEBUG nova.compute.manager [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Instance network_info: |[{"id": "943c04a5-7eba-46db-ab4f-93f7f84be14e", "address": "fa:16:3e:56:de:39", "network": {"id": "7635ca79-1de7-4ba3-b347-1e0b66d7eb56", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1574772814-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8e9d6e6ceac4520add365429c4eb4d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d0c6fd7-3cc9-4818-9475-8f15900394cc", "external-id": "nsx-vlan-transportzone-317", "segmentation_id": 317, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap943c04a5-7e", "ovs_interfaceid": "943c04a5-7eba-46db-ab4f-93f7f84be14e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 802.388964] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:56:de:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d0c6fd7-3cc9-4818-9475-8f15900394cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '943c04a5-7eba-46db-ab4f-93f7f84be14e', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 802.396362] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Creating folder: Project (c8e9d6e6ceac4520add365429c4eb4d2). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 802.396874] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5f9ee064-a12d-4aaa-bb8f-1ecb7b4cef84 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.407228] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Created folder: Project (c8e9d6e6ceac4520add365429c4eb4d2) in parent group-v274258. [ 802.407417] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Creating folder: Instances. Parent ref: group-v274312. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 802.407637] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-10f0f76d-fdf3-4a2a-9f2f-1f0443b20982 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.415656] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Created folder: Instances in parent group-v274312. [ 802.415883] env[61867]: DEBUG oslo.service.loopingcall [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 802.416091] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 802.416301] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-45ef04a4-c6b3-43cd-b597-799b7dca2294 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.435498] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 802.435498] env[61867]: value = "task-1276498" [ 802.435498] env[61867]: _type = "Task" [ 802.435498] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.441178] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276495, 'name': ReconfigVM_Task, 'duration_secs': 0.26871} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.441771] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Reconfigured VM instance instance-0000003d to attach disk [datastore2] 80a703b3-d692-4023-a73b-dba980a94dff/80a703b3-d692-4023-a73b-dba980a94dff.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 802.442383] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-66ce6b90-36ba-46ea-bef4-9cdd8dc17718 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.447197] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276498, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.451663] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 802.451663] env[61867]: value = "task-1276499" [ 802.451663] env[61867]: _type = "Task" [ 802.451663] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.459948] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276499, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.511302] env[61867]: DEBUG nova.scheduler.client.report [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 802.596763] env[61867]: INFO nova.compute.manager [-] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Took 1.36 seconds to deallocate network for instance. [ 802.608321] env[61867]: DEBUG nova.compute.manager [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 802.616382] env[61867]: DEBUG nova.network.neutron [req-2b07905b-e720-4522-95a0-2f06f9e0f10f req-e2d8fef6-85fc-4c3c-9cee-117546cdab74 service nova] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.707224] env[61867]: DEBUG nova.network.neutron [req-2b07905b-e720-4522-95a0-2f06f9e0f10f req-e2d8fef6-85fc-4c3c-9cee-117546cdab74 service nova] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.855125] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "be825f62-cb56-4647-be1c-1605536fdc89" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.855402] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "be825f62-cb56-4647-be1c-1605536fdc89" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.855610] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "be825f62-cb56-4647-be1c-1605536fdc89-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.855798] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "be825f62-cb56-4647-be1c-1605536fdc89-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.855970] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "be825f62-cb56-4647-be1c-1605536fdc89-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.858216] env[61867]: INFO nova.compute.manager [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Terminating instance [ 802.859954] env[61867]: DEBUG nova.compute.manager [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 802.860230] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 802.861178] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-339d035a-0af0-4b99-a941-1acc7a542c68 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.868768] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 802.868987] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-458f708a-2e98-433b-8c7d-5676efddb8f2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.875872] env[61867]: DEBUG oslo_vmware.api [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 802.875872] env[61867]: value = "task-1276500" [ 802.875872] env[61867]: _type = "Task" [ 802.875872] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.885164] env[61867]: DEBUG oslo_vmware.api [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276500, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.948343] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276498, 'name': CreateVM_Task, 'duration_secs': 0.368529} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.948669] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 802.949558] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.949730] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.950225] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 802.950559] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea0014e2-ba80-40bd-a359-9dbd9d66d16f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.958673] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 802.958673] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5221f6bb-b166-e12b-3d25-f229458b6975" [ 802.958673] env[61867]: _type = "Task" [ 802.958673] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.962297] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276499, 'name': Rename_Task, 'duration_secs': 0.193921} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.965383] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 802.965638] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-625e5ead-307a-44d9-9654-723cc0699e00 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.972718] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5221f6bb-b166-e12b-3d25-f229458b6975, 'name': SearchDatastore_Task, 'duration_secs': 0.009752} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.973883] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.974171] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 802.974433] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.974586] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.974765] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 802.975446] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 802.975446] env[61867]: value = "task-1276501" [ 802.975446] env[61867]: _type = "Task" [ 802.975446] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.975646] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-13d2090c-6dbd-4ab2-b80a-94086c07c4cb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.985923] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276501, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.987063] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 802.987063] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 802.987317] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e73f1ffc-cd13-4ffc-9bb6-7901366e3e2d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.991759] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 802.991759] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b5333b-cc53-5fa1-4fb8-dd497b769d66" [ 802.991759] env[61867]: _type = "Task" [ 802.991759] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.999637] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b5333b-cc53-5fa1-4fb8-dd497b769d66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.016862] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.858s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.019311] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.988s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.020842] env[61867]: INFO nova.compute.claims [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 803.040815] env[61867]: INFO nova.scheduler.client.report [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Deleted allocations for instance 81ada1e9-b100-44fd-8119-7ef8695e601f [ 803.104199] env[61867]: DEBUG oslo_concurrency.lockutils [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.129711] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.210158] env[61867]: DEBUG oslo_concurrency.lockutils [req-2b07905b-e720-4522-95a0-2f06f9e0f10f req-e2d8fef6-85fc-4c3c-9cee-117546cdab74 service nova] Releasing lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.211382] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquired lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.211382] env[61867]: DEBUG nova.network.neutron [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 803.385526] env[61867]: DEBUG oslo_vmware.api [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276500, 'name': PowerOffVM_Task, 'duration_secs': 0.182152} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.385746] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 803.385918] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 803.386176] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-775091b2-7d9a-4443-b135-26acf606665d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.452418] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 803.452780] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 803.452826] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleting the datastore file [datastore2] be825f62-cb56-4647-be1c-1605536fdc89 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 803.453171] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d38adab4-dae6-4ed8-8e20-9b20699eaf38 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.460242] env[61867]: DEBUG oslo_vmware.api [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 803.460242] env[61867]: value = "task-1276503" [ 803.460242] env[61867]: _type = "Task" [ 803.460242] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.468235] env[61867]: DEBUG oslo_vmware.api [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276503, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.485897] env[61867]: DEBUG oslo_vmware.api [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276501, 'name': PowerOnVM_Task, 'duration_secs': 0.474915} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.486181] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 803.486397] env[61867]: INFO nova.compute.manager [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Took 6.94 seconds to spawn the instance on the hypervisor. [ 803.486594] env[61867]: DEBUG nova.compute.manager [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 803.487360] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e23cbfd-ea79-48d3-8ec9-c6a8306f45a7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.503376] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b5333b-cc53-5fa1-4fb8-dd497b769d66, 'name': SearchDatastore_Task, 'duration_secs': 0.008896} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.504070] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e070e26-d6fc-44c4-abba-85a635b8e7ff {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.508713] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 803.508713] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5249f138-ba80-b549-fcc8-e0677734cd89" [ 803.508713] env[61867]: _type = "Task" [ 803.508713] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.516354] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5249f138-ba80-b549-fcc8-e0677734cd89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.549278] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9989a74f-6b4e-4ee3-b3b5-4f88fdce68c7 tempest-MultipleCreateTestJSON-1919498429 tempest-MultipleCreateTestJSON-1919498429-project-member] Lock "81ada1e9-b100-44fd-8119-7ef8695e601f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.199s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.763340] env[61867]: DEBUG nova.network.neutron [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 803.973604] env[61867]: DEBUG oslo_vmware.api [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276503, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.444448} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.973604] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 803.973604] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 803.973604] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 803.973604] env[61867]: INFO nova.compute.manager [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Took 1.11 seconds to destroy the instance on the hypervisor. [ 803.973856] env[61867]: DEBUG oslo.service.loopingcall [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 803.973856] env[61867]: DEBUG nova.compute.manager [-] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 803.973856] env[61867]: DEBUG nova.network.neutron [-] [instance: be825f62-cb56-4647-be1c-1605536fdc89] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 804.005797] env[61867]: INFO nova.compute.manager [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Took 29.33 seconds to build instance. [ 804.008108] env[61867]: DEBUG nova.network.neutron [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Updating instance_info_cache with network_info: [{"id": "79dddbac-ec4f-4cd9-8436-a3600957e598", "address": "fa:16:3e:b8:10:25", "network": {"id": "2d4ab52c-144f-45a1-9d4f-afd85fc30404", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.63", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "84601e35c8e8487cb78fc16a2536a4c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d47d5e1d-e66d-4f2c-83e6-d5e78c2b767d", "external-id": "nsx-vlan-transportzone-109", "segmentation_id": 109, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79dddbac-ec", "ovs_interfaceid": "79dddbac-ec4f-4cd9-8436-a3600957e598", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.019670] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5249f138-ba80-b549-fcc8-e0677734cd89, 'name': SearchDatastore_Task, 'duration_secs': 0.009982} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.020159] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.020693] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 1df8427c-e75d-4b60-a92a-b5ba76b67081/1df8427c-e75d-4b60-a92a-b5ba76b67081.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 804.021346] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-568eb638-ba75-44cb-ab77-a8d49947c373 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.031640] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 804.031640] env[61867]: value = "task-1276504" [ 804.031640] env[61867]: _type = "Task" [ 804.031640] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.044492] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276504, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.308242] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3694a8af-b732-4ec9-a2f1-24faaad3a577 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.317154] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15419f43-8a4f-49fc-8ed7-afcab23bc55d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.351083] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddadfb9b-54ca-4567-8e44-af4f396469e5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.360277] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6993e0bc-b354-461a-861b-502599cf0833 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.375469] env[61867]: DEBUG nova.compute.provider_tree [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.476451] env[61867]: DEBUG nova.compute.manager [req-3a99771f-7a7f-40be-ab22-f5956df06e29 req-6c50faa1-ad9a-43cd-b7e8-ba5767a9f181 service nova] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Received event network-vif-deleted-d9ab90ee-f219-4de7-af2a-3db270f0f2c6 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.477242] env[61867]: INFO nova.compute.manager [req-3a99771f-7a7f-40be-ab22-f5956df06e29 req-6c50faa1-ad9a-43cd-b7e8-ba5767a9f181 service nova] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Neutron deleted interface d9ab90ee-f219-4de7-af2a-3db270f0f2c6; detaching it from the instance and deleting it from the info cache [ 804.477242] env[61867]: DEBUG nova.network.neutron [req-3a99771f-7a7f-40be-ab22-f5956df06e29 req-6c50faa1-ad9a-43cd-b7e8-ba5767a9f181 service nova] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.511299] env[61867]: DEBUG oslo_concurrency.lockutils [None req-381a476a-34e4-4b20-ab7c-037c975171ea tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "80a703b3-d692-4023-a73b-dba980a94dff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.166s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.514689] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Releasing lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.514789] env[61867]: DEBUG nova.compute.manager [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Instance network_info: |[{"id": "79dddbac-ec4f-4cd9-8436-a3600957e598", "address": "fa:16:3e:b8:10:25", "network": {"id": "2d4ab52c-144f-45a1-9d4f-afd85fc30404", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.63", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "84601e35c8e8487cb78fc16a2536a4c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d47d5e1d-e66d-4f2c-83e6-d5e78c2b767d", "external-id": "nsx-vlan-transportzone-109", "segmentation_id": 109, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79dddbac-ec", "ovs_interfaceid": "79dddbac-ec4f-4cd9-8436-a3600957e598", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 804.515206] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:10:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd47d5e1d-e66d-4f2c-83e6-d5e78c2b767d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '79dddbac-ec4f-4cd9-8436-a3600957e598', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 804.522657] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Creating folder: Project (2b64ff57036a492eb2555f33e0af664e). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 804.523376] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dde8a138-96a1-4ed2-bc0d-754b8f7eb867 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.533047] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Created folder: Project (2b64ff57036a492eb2555f33e0af664e) in parent group-v274258. [ 804.533243] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Creating folder: Instances. Parent ref: group-v274315. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 804.536559] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1cd097c6-1c69-4f14-bed8-c00f1f67fe78 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.542823] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276504, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492574} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.543069] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 1df8427c-e75d-4b60-a92a-b5ba76b67081/1df8427c-e75d-4b60-a92a-b5ba76b67081.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 804.543284] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 804.543506] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c148b1da-da93-4e33-93f5-e311cba57684 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.546140] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Created folder: Instances in parent group-v274315. [ 804.546360] env[61867]: DEBUG oslo.service.loopingcall [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.547327] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 804.547588] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2fd1eea2-2317-45df-8cff-248c2771892b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.563371] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 804.563371] env[61867]: value = "task-1276507" [ 804.563371] env[61867]: _type = "Task" [ 804.563371] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.568472] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 804.568472] env[61867]: value = "task-1276508" [ 804.568472] env[61867]: _type = "Task" [ 804.568472] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.571641] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276507, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.578730] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276508, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.878453] env[61867]: DEBUG nova.scheduler.client.report [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.895648] env[61867]: DEBUG nova.network.neutron [-] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.980685] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a97d6106-6ee0-4745-8b96-e12c51a262a3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.989363] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a289b033-4c0d-4d68-bc07-f7c3434004ca {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.013650] env[61867]: DEBUG nova.compute.manager [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 805.024836] env[61867]: DEBUG nova.compute.manager [req-3a99771f-7a7f-40be-ab22-f5956df06e29 req-6c50faa1-ad9a-43cd-b7e8-ba5767a9f181 service nova] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Detach interface failed, port_id=d9ab90ee-f219-4de7-af2a-3db270f0f2c6, reason: Instance be825f62-cb56-4647-be1c-1605536fdc89 could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 805.074492] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276507, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065435} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.079367] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 805.080472] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc400baa-c970-477e-a893-c4f1ed5196a1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.091548] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276508, 'name': CreateVM_Task, 'duration_secs': 0.406703} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.091548] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 805.093167] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.093167] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.093167] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 805.106544] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d762414-2c82-4d92-9def-612d808914c7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.116917] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 1df8427c-e75d-4b60-a92a-b5ba76b67081/1df8427c-e75d-4b60-a92a-b5ba76b67081.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 805.119142] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68d55cec-7a79-4109-b4a0-71ef9fe993d9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.135630] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 805.135630] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5297a2c9-137c-8cd1-7a2c-5c54604ad4c3" [ 805.135630] env[61867]: _type = "Task" [ 805.135630] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.139958] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 805.139958] env[61867]: value = "task-1276509" [ 805.139958] env[61867]: _type = "Task" [ 805.139958] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.146602] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5297a2c9-137c-8cd1-7a2c-5c54604ad4c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.152299] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276509, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.384012] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.384573] env[61867]: DEBUG nova.compute.manager [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 805.387345] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 24.331s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.401655] env[61867]: INFO nova.compute.manager [-] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Took 1.43 seconds to deallocate network for instance. [ 805.541986] env[61867]: DEBUG oslo_concurrency.lockutils [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.593433] env[61867]: DEBUG nova.compute.manager [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 805.594355] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d129d4b5-914b-4352-b705-dc3e506d18f2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.651460] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5297a2c9-137c-8cd1-7a2c-5c54604ad4c3, 'name': SearchDatastore_Task, 'duration_secs': 0.030854} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.654761] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.655007] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 805.655247] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.655397] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.655573] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 805.655827] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276509, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.656042] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-58018ec3-d099-4bd0-880d-af261e32fdbc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.668141] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 805.668322] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 805.669032] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d1d0f0e-b5fc-4708-8efd-360d5d7448df {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.674273] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 805.674273] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5278ddb0-cd61-0e70-da41-be899d8c0baa" [ 805.674273] env[61867]: _type = "Task" [ 805.674273] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.681566] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5278ddb0-cd61-0e70-da41-be899d8c0baa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.891380] env[61867]: DEBUG nova.compute.utils [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 805.898139] env[61867]: DEBUG nova.compute.manager [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 805.898256] env[61867]: DEBUG nova.network.neutron [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 805.908012] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.950403] env[61867]: DEBUG nova.policy [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '60a1d41ef1a74be5b47aa17478a679fe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a179f33eccc54c5784b1b1acc1795ee4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 806.106370] env[61867]: INFO nova.compute.manager [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] instance snapshotting [ 806.109307] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6aa09f-bdeb-4bf2-8cb7-d4330222265a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.135252] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64507635-ad5d-45e9-b553-c6c2ee203b16 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.155551] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276509, 'name': ReconfigVM_Task, 'duration_secs': 0.953659} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.155551] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 1df8427c-e75d-4b60-a92a-b5ba76b67081/1df8427c-e75d-4b60-a92a-b5ba76b67081.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 806.159144] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a56de216-3532-48fa-9daf-ecf80ba905e6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.165199] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 806.165199] env[61867]: value = "task-1276510" [ 806.165199] env[61867]: _type = "Task" [ 806.165199] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.174773] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276510, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.184029] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5278ddb0-cd61-0e70-da41-be899d8c0baa, 'name': SearchDatastore_Task, 'duration_secs': 0.086339} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.184997] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e48e04ab-224c-4e7c-8504-be13e07d1277 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.191529] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 806.191529] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b6fb59-3b8f-2c4f-e5f1-214ae18125bb" [ 806.191529] env[61867]: _type = "Task" [ 806.191529] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.200051] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b6fb59-3b8f-2c4f-e5f1-214ae18125bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.260848] env[61867]: DEBUG nova.network.neutron [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Successfully created port: 5a9bc008-86f3-4487-898e-f6d09ec8f111 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 806.398453] env[61867]: DEBUG nova.compute.manager [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 806.425147] env[61867]: WARNING nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance b1ec32c5-0642-4dca-ad43-05a5172d04d7 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 806.425338] env[61867]: WARNING nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance c7203cea-2c2d-4a97-8ff4-3d33cd8443f7 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 806.425474] env[61867]: WARNING nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance f0b6273f-4177-44a3-8dd1-f65faeb07539 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 806.425604] env[61867]: WARNING nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 905ee661-d5bd-4121-9bd0-fe68bdabd2fd is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 806.425729] env[61867]: WARNING nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 3022ea79-7c27-40a4-80d6-1a86b354917a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 806.425888] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance be825f62-cb56-4647-be1c-1605536fdc89 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.425964] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 80a703b3-d692-4023-a73b-dba980a94dff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.426095] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 1df8427c-e75d-4b60-a92a-b5ba76b67081 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.426214] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 81e70c36-04ea-450c-9383-53ef069d1c46 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.426348] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 4653ce96-257e-46d8-8c3d-85c03380213c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.651355] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Creating Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 806.651709] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c41fd554-efae-4735-9f8c-3ae9b2ba0dde {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.662497] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 806.662497] env[61867]: value = "task-1276511" [ 806.662497] env[61867]: _type = "Task" [ 806.662497] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.674941] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276511, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.677294] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276510, 'name': Rename_Task, 'duration_secs': 0.291513} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.677664] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 806.677951] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d2cc115-acf0-4da4-a381-fa1112f507ad {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.683929] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 806.683929] env[61867]: value = "task-1276512" [ 806.683929] env[61867]: _type = "Task" [ 806.683929] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.694155] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276512, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.701720] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b6fb59-3b8f-2c4f-e5f1-214ae18125bb, 'name': SearchDatastore_Task, 'duration_secs': 0.009804} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.701963] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.702243] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 81e70c36-04ea-450c-9383-53ef069d1c46/81e70c36-04ea-450c-9383-53ef069d1c46.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 806.702489] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9e5c1e86-6f4c-4878-a24d-479bbf31ec2a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.708256] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 806.708256] env[61867]: value = "task-1276513" [ 806.708256] env[61867]: _type = "Task" [ 806.708256] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.715299] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276513, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.929334] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 66518b2a-0242-438b-ba9f-d57c07a1165c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 807.173243] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276511, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.193308] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276512, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.219257] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276513, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446909} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.219533] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 81e70c36-04ea-450c-9383-53ef069d1c46/81e70c36-04ea-450c-9383-53ef069d1c46.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 807.219797] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.220253] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ffff5472-4b3a-4b54-a32b-099e31f04bb0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.228023] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 807.228023] env[61867]: value = "task-1276514" [ 807.228023] env[61867]: _type = "Task" [ 807.228023] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.235751] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276514, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.410300] env[61867]: DEBUG nova.compute.manager [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 807.434396] env[61867]: DEBUG nova.virt.hardware [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 807.434643] env[61867]: DEBUG nova.virt.hardware [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 807.434808] env[61867]: DEBUG nova.virt.hardware [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 807.434992] env[61867]: DEBUG nova.virt.hardware [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 807.435160] env[61867]: DEBUG nova.virt.hardware [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 807.435307] env[61867]: DEBUG nova.virt.hardware [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 807.435514] env[61867]: DEBUG nova.virt.hardware [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 807.435665] env[61867]: DEBUG nova.virt.hardware [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 807.435828] env[61867]: DEBUG nova.virt.hardware [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 807.435991] env[61867]: DEBUG nova.virt.hardware [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 807.436264] env[61867]: DEBUG nova.virt.hardware [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 807.436956] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance abb41c0c-6d0d-4147-a4af-554ab7d9e921 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 807.439507] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a60336e6-7821-48cc-bba1-e1e11fcae66e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.448338] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65430add-fce0-4737-b17b-d4459948dcc2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.675712] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276511, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.695240] env[61867]: DEBUG oslo_vmware.api [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276512, 'name': PowerOnVM_Task, 'duration_secs': 0.698358} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.695423] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 807.695691] env[61867]: INFO nova.compute.manager [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Took 8.75 seconds to spawn the instance on the hypervisor. [ 807.695902] env[61867]: DEBUG nova.compute.manager [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 807.696932] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab00740-6f8a-45de-a7e4-3fb418e641c0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.737022] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276514, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068265} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.737783] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 807.740158] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf573960-8328-4811-9af1-47e3456260db {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.763660] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 81e70c36-04ea-450c-9383-53ef069d1c46/81e70c36-04ea-450c-9383-53ef069d1c46.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 807.764918] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59c7c9bf-178c-483a-9883-4d00021f7faa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.784373] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 807.784373] env[61867]: value = "task-1276515" [ 807.784373] env[61867]: _type = "Task" [ 807.784373] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.796141] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276515, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.819556] env[61867]: DEBUG nova.network.neutron [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Successfully updated port: 5a9bc008-86f3-4487-898e-f6d09ec8f111 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 807.849931] env[61867]: DEBUG nova.compute.manager [req-827e9a79-fb66-4b38-acc5-4d8623b411f4 req-1342d09a-d14a-4cfc-951d-1cdb49f7d1b0 service nova] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Received event network-vif-plugged-5a9bc008-86f3-4487-898e-f6d09ec8f111 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 807.850188] env[61867]: DEBUG oslo_concurrency.lockutils [req-827e9a79-fb66-4b38-acc5-4d8623b411f4 req-1342d09a-d14a-4cfc-951d-1cdb49f7d1b0 service nova] Acquiring lock "4653ce96-257e-46d8-8c3d-85c03380213c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.850329] env[61867]: DEBUG oslo_concurrency.lockutils [req-827e9a79-fb66-4b38-acc5-4d8623b411f4 req-1342d09a-d14a-4cfc-951d-1cdb49f7d1b0 service nova] Lock "4653ce96-257e-46d8-8c3d-85c03380213c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.850503] env[61867]: DEBUG oslo_concurrency.lockutils [req-827e9a79-fb66-4b38-acc5-4d8623b411f4 req-1342d09a-d14a-4cfc-951d-1cdb49f7d1b0 service nova] Lock "4653ce96-257e-46d8-8c3d-85c03380213c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.850674] env[61867]: DEBUG nova.compute.manager [req-827e9a79-fb66-4b38-acc5-4d8623b411f4 req-1342d09a-d14a-4cfc-951d-1cdb49f7d1b0 service nova] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] No waiting events found dispatching network-vif-plugged-5a9bc008-86f3-4487-898e-f6d09ec8f111 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 807.850834] env[61867]: WARNING nova.compute.manager [req-827e9a79-fb66-4b38-acc5-4d8623b411f4 req-1342d09a-d14a-4cfc-951d-1cdb49f7d1b0 service nova] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Received unexpected event network-vif-plugged-5a9bc008-86f3-4487-898e-f6d09ec8f111 for instance with vm_state building and task_state spawning. [ 807.945160] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 6e41989e-b8fa-4009-af1e-1ce859b329a1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 808.175710] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276511, 'name': CreateSnapshot_Task, 'duration_secs': 1.063197} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.176135] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Created Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 808.177083] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b9a443-3715-4177-b1d4-bbfc74f6f375 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.215017] env[61867]: INFO nova.compute.manager [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Took 31.93 seconds to build instance. [ 808.299419] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276515, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.321280] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Acquiring lock "refresh_cache-4653ce96-257e-46d8-8c3d-85c03380213c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.321434] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Acquired lock "refresh_cache-4653ce96-257e-46d8-8c3d-85c03380213c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.321585] env[61867]: DEBUG nova.network.neutron [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 808.447561] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 8a83f4f2-58eb-473e-9b1e-32ce633554f9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 808.700063] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Creating linked-clone VM from snapshot {{(pid=61867) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 808.700063] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6535f8cb-218d-4a04-b532-3a1b4617b0c9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.707305] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 808.707305] env[61867]: value = "task-1276516" [ 808.707305] env[61867]: _type = "Task" [ 808.707305] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.716705] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276516, 'name': CloneVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.718293] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a0face50-f162-440f-aa47-fc3244b5de55 tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "1df8427c-e75d-4b60-a92a-b5ba76b67081" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.782s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.797336] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276515, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.939419] env[61867]: DEBUG nova.network.neutron [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.956788] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 6dda389d-a8c5-4e0e-87a5-4065e24c034e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 809.214249] env[61867]: DEBUG nova.network.neutron [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Updating instance_info_cache with network_info: [{"id": "5a9bc008-86f3-4487-898e-f6d09ec8f111", "address": "fa:16:3e:60:40:2a", "network": {"id": "d607b3c4-79c8-42ba-88ba-9c117903a773", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1791928491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a179f33eccc54c5784b1b1acc1795ee4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "489b2441-7132-4942-8b61-49cf0ad4400e", "external-id": "nsx-vlan-transportzone-971", "segmentation_id": 971, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a9bc008-86", "ovs_interfaceid": "5a9bc008-86f3-4487-898e-f6d09ec8f111", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.222667] env[61867]: DEBUG nova.compute.manager [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 809.232266] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276516, 'name': CloneVM_Task} progress is 93%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.306461] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276515, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.459405] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 9aab8852-addb-49e6-a59b-fa9bffc7733b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 809.718411] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Releasing lock "refresh_cache-4653ce96-257e-46d8-8c3d-85c03380213c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.718760] env[61867]: DEBUG nova.compute.manager [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Instance network_info: |[{"id": "5a9bc008-86f3-4487-898e-f6d09ec8f111", "address": "fa:16:3e:60:40:2a", "network": {"id": "d607b3c4-79c8-42ba-88ba-9c117903a773", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1791928491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a179f33eccc54c5784b1b1acc1795ee4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "489b2441-7132-4942-8b61-49cf0ad4400e", "external-id": "nsx-vlan-transportzone-971", "segmentation_id": 971, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a9bc008-86", "ovs_interfaceid": "5a9bc008-86f3-4487-898e-f6d09ec8f111", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 809.719024] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276516, 'name': CloneVM_Task} progress is 93%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.719459] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:40:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '489b2441-7132-4942-8b61-49cf0ad4400e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5a9bc008-86f3-4487-898e-f6d09ec8f111', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 809.726920] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Creating folder: Project (a179f33eccc54c5784b1b1acc1795ee4). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 809.727223] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cc8aac25-438d-4479-a339-30e3a259a00c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.740492] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Created folder: Project (a179f33eccc54c5784b1b1acc1795ee4) in parent group-v274258. [ 809.740690] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Creating folder: Instances. Parent ref: group-v274320. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 809.740929] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5aefd116-d06e-48d8-b3ff-03ce1c5c3f5b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.750417] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Created folder: Instances in parent group-v274320. [ 809.750665] env[61867]: DEBUG oslo.service.loopingcall [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.750855] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 809.751083] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d5411f2b-b9c4-4903-85ad-1ebc15bbbf40 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.766218] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.771339] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 809.771339] env[61867]: value = "task-1276519" [ 809.771339] env[61867]: _type = "Task" [ 809.771339] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.783439] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276519, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.795765] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Acquiring lock "410a2819-ea27-4613-9f2c-279f9e82a4b3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.796018] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Lock "410a2819-ea27-4613-9f2c-279f9e82a4b3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.800275] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276515, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.881414] env[61867]: DEBUG nova.compute.manager [req-2f722623-0b80-4661-8955-29293fb23871 req-86fa0c45-1bdc-4aa9-be81-babb17b7d040 service nova] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Received event network-changed-5a9bc008-86f3-4487-898e-f6d09ec8f111 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.881625] env[61867]: DEBUG nova.compute.manager [req-2f722623-0b80-4661-8955-29293fb23871 req-86fa0c45-1bdc-4aa9-be81-babb17b7d040 service nova] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Refreshing instance network info cache due to event network-changed-5a9bc008-86f3-4487-898e-f6d09ec8f111. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 809.881839] env[61867]: DEBUG oslo_concurrency.lockutils [req-2f722623-0b80-4661-8955-29293fb23871 req-86fa0c45-1bdc-4aa9-be81-babb17b7d040 service nova] Acquiring lock "refresh_cache-4653ce96-257e-46d8-8c3d-85c03380213c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.882018] env[61867]: DEBUG oslo_concurrency.lockutils [req-2f722623-0b80-4661-8955-29293fb23871 req-86fa0c45-1bdc-4aa9-be81-babb17b7d040 service nova] Acquired lock "refresh_cache-4653ce96-257e-46d8-8c3d-85c03380213c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.882425] env[61867]: DEBUG nova.network.neutron [req-2f722623-0b80-4661-8955-29293fb23871 req-86fa0c45-1bdc-4aa9-be81-babb17b7d040 service nova] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Refreshing network info cache for port 5a9bc008-86f3-4487-898e-f6d09ec8f111 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 809.962722] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance ad699b24-d01d-4d7a-815f-c6b10286012d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 810.222975] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276516, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.282904] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276519, 'name': CreateVM_Task, 'duration_secs': 0.316004} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.283114] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 810.284293] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.284579] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.284795] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 810.285218] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-826127df-6084-4fde-afb9-139d5e227f48 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.292773] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Waiting for the task: (returnval){ [ 810.292773] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528e99f7-9f42-43a5-3ad6-373d054eadbd" [ 810.292773] env[61867]: _type = "Task" [ 810.292773] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.299606] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276515, 'name': ReconfigVM_Task, 'duration_secs': 2.29768} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.300279] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 81e70c36-04ea-450c-9383-53ef069d1c46/81e70c36-04ea-450c-9383-53ef069d1c46.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 810.300927] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5d6720c3-1925-4df8-9808-95f4dc396684 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.306334] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528e99f7-9f42-43a5-3ad6-373d054eadbd, 'name': SearchDatastore_Task, 'duration_secs': 0.008929} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.306992] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.307350] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 810.307604] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.307774] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.307978] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 810.308296] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b5695f16-2b6d-4179-91a7-1b2a20ca7a07 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.311692] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 810.311692] env[61867]: value = "task-1276520" [ 810.311692] env[61867]: _type = "Task" [ 810.311692] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.316461] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 810.316664] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 810.317782] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d03a95a6-f225-4413-9b15-85d1f3337acf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.323250] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276520, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.326374] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Waiting for the task: (returnval){ [ 810.326374] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f5d3c9-22b0-0c2d-8074-9452b8c57980" [ 810.326374] env[61867]: _type = "Task" [ 810.326374] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.335361] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f5d3c9-22b0-0c2d-8074-9452b8c57980, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.468471] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance ca2f6d18-f773-4875-83de-2f2be912c2f8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 810.469028] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 810.469222] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 810.652846] env[61867]: DEBUG nova.network.neutron [req-2f722623-0b80-4661-8955-29293fb23871 req-86fa0c45-1bdc-4aa9-be81-babb17b7d040 service nova] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Updated VIF entry in instance network info cache for port 5a9bc008-86f3-4487-898e-f6d09ec8f111. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 810.654223] env[61867]: DEBUG nova.network.neutron [req-2f722623-0b80-4661-8955-29293fb23871 req-86fa0c45-1bdc-4aa9-be81-babb17b7d040 service nova] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Updating instance_info_cache with network_info: [{"id": "5a9bc008-86f3-4487-898e-f6d09ec8f111", "address": "fa:16:3e:60:40:2a", "network": {"id": "d607b3c4-79c8-42ba-88ba-9c117903a773", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1791928491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a179f33eccc54c5784b1b1acc1795ee4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "489b2441-7132-4942-8b61-49cf0ad4400e", "external-id": "nsx-vlan-transportzone-971", "segmentation_id": 971, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a9bc008-86", "ovs_interfaceid": "5a9bc008-86f3-4487-898e-f6d09ec8f111", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.720262] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276516, 'name': CloneVM_Task} progress is 95%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.774556] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-834c479f-7bfa-471e-ae70-3ec5a3338467 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.781810] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2307b265-e0c5-4860-9187-a16be9dc309c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.813135] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f002e8ec-0934-4e2a-a6fc-fc0051e33a78 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.827233] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aba250e-2c82-4ed0-9010-50df7c276571 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.830773] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276520, 'name': Rename_Task, 'duration_secs': 0.149447} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.831101] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 810.834139] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2b93200-01dc-4e55-b400-9451da3f4c07 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.843423] env[61867]: DEBUG nova.compute.provider_tree [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.848306] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 810.848306] env[61867]: value = "task-1276521" [ 810.848306] env[61867]: _type = "Task" [ 810.848306] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.848516] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f5d3c9-22b0-0c2d-8074-9452b8c57980, 'name': SearchDatastore_Task, 'duration_secs': 0.011063} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.851917] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b97fa6b-6f14-4618-9e4a-b672985c92c0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.859668] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Waiting for the task: (returnval){ [ 810.859668] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b872cf-9aef-dba4-b905-bf87cca35ed0" [ 810.859668] env[61867]: _type = "Task" [ 810.859668] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.867101] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b872cf-9aef-dba4-b905-bf87cca35ed0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.157573] env[61867]: DEBUG oslo_concurrency.lockutils [req-2f722623-0b80-4661-8955-29293fb23871 req-86fa0c45-1bdc-4aa9-be81-babb17b7d040 service nova] Releasing lock "refresh_cache-4653ce96-257e-46d8-8c3d-85c03380213c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.157876] env[61867]: DEBUG nova.compute.manager [req-2f722623-0b80-4661-8955-29293fb23871 req-86fa0c45-1bdc-4aa9-be81-babb17b7d040 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Received event network-changed-943c04a5-7eba-46db-ab4f-93f7f84be14e {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.158074] env[61867]: DEBUG nova.compute.manager [req-2f722623-0b80-4661-8955-29293fb23871 req-86fa0c45-1bdc-4aa9-be81-babb17b7d040 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Refreshing instance network info cache due to event network-changed-943c04a5-7eba-46db-ab4f-93f7f84be14e. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 811.158337] env[61867]: DEBUG oslo_concurrency.lockutils [req-2f722623-0b80-4661-8955-29293fb23871 req-86fa0c45-1bdc-4aa9-be81-babb17b7d040 service nova] Acquiring lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.158589] env[61867]: DEBUG oslo_concurrency.lockutils [req-2f722623-0b80-4661-8955-29293fb23871 req-86fa0c45-1bdc-4aa9-be81-babb17b7d040 service nova] Acquired lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.158797] env[61867]: DEBUG nova.network.neutron [req-2f722623-0b80-4661-8955-29293fb23871 req-86fa0c45-1bdc-4aa9-be81-babb17b7d040 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Refreshing network info cache for port 943c04a5-7eba-46db-ab4f-93f7f84be14e {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 811.219912] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276516, 'name': CloneVM_Task, 'duration_secs': 2.121664} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.220242] env[61867]: INFO nova.virt.vmwareapi.vmops [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Created linked-clone VM from snapshot [ 811.220972] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6fb569f-eabf-4b91-8513-f1d472199a44 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.228090] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Uploading image 420f5356-3ba1-4114-af3a-5be2a65b82cb {{(pid=61867) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 811.248232] env[61867]: DEBUG oslo_vmware.rw_handles [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 811.248232] env[61867]: value = "vm-274319" [ 811.248232] env[61867]: _type = "VirtualMachine" [ 811.248232] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 811.248495] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f3ee2a2c-b191-4b04-9710-b0dac62e3e9f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.255731] env[61867]: DEBUG oslo_vmware.rw_handles [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lease: (returnval){ [ 811.255731] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5284408f-843c-6fc6-7d56-3e9694dc3fcc" [ 811.255731] env[61867]: _type = "HttpNfcLease" [ 811.255731] env[61867]: } obtained for exporting VM: (result){ [ 811.255731] env[61867]: value = "vm-274319" [ 811.255731] env[61867]: _type = "VirtualMachine" [ 811.255731] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 811.255964] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the lease: (returnval){ [ 811.255964] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5284408f-843c-6fc6-7d56-3e9694dc3fcc" [ 811.255964] env[61867]: _type = "HttpNfcLease" [ 811.255964] env[61867]: } to be ready. {{(pid=61867) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 811.262412] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 811.262412] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5284408f-843c-6fc6-7d56-3e9694dc3fcc" [ 811.262412] env[61867]: _type = "HttpNfcLease" [ 811.262412] env[61867]: } is initializing. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 811.350136] env[61867]: DEBUG nova.scheduler.client.report [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.364312] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276521, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.374139] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b872cf-9aef-dba4-b905-bf87cca35ed0, 'name': SearchDatastore_Task, 'duration_secs': 0.013174} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.374439] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.374829] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 4653ce96-257e-46d8-8c3d-85c03380213c/4653ce96-257e-46d8-8c3d-85c03380213c.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 811.374960] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-983ef4ae-1337-4570-84d8-d90e0dc66639 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.381800] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Waiting for the task: (returnval){ [ 811.381800] env[61867]: value = "task-1276523" [ 811.381800] env[61867]: _type = "Task" [ 811.381800] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.391248] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': task-1276523, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.765154] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 811.765154] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5284408f-843c-6fc6-7d56-3e9694dc3fcc" [ 811.765154] env[61867]: _type = "HttpNfcLease" [ 811.765154] env[61867]: } is ready. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 811.766709] env[61867]: DEBUG oslo_vmware.rw_handles [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 811.766709] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5284408f-843c-6fc6-7d56-3e9694dc3fcc" [ 811.766709] env[61867]: _type = "HttpNfcLease" [ 811.766709] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 811.766709] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d02e69b5-cf4a-4cf9-8199-79308c1bfcd6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.775087] env[61867]: DEBUG oslo_vmware.rw_handles [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529c83e1-d237-3f7b-55b1-69622459ec92/disk-0.vmdk from lease info. {{(pid=61867) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 811.775333] env[61867]: DEBUG oslo_vmware.rw_handles [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529c83e1-d237-3f7b-55b1-69622459ec92/disk-0.vmdk for reading. {{(pid=61867) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 811.858616] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61867) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 811.858850] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.472s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.859153] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.005s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.861989] env[61867]: INFO nova.compute.claims [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 811.871247] env[61867]: DEBUG oslo_vmware.api [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276521, 'name': PowerOnVM_Task, 'duration_secs': 0.91426} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.871637] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 811.871849] env[61867]: INFO nova.compute.manager [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Took 10.60 seconds to spawn the instance on the hypervisor. [ 811.872040] env[61867]: DEBUG nova.compute.manager [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 811.872949] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f3ac72e-efc1-4e77-a277-ee1a6589f148 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.891636] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': task-1276523, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.977249] env[61867]: DEBUG nova.network.neutron [req-2f722623-0b80-4661-8955-29293fb23871 req-86fa0c45-1bdc-4aa9-be81-babb17b7d040 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Updated VIF entry in instance network info cache for port 943c04a5-7eba-46db-ab4f-93f7f84be14e. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 811.977494] env[61867]: DEBUG nova.network.neutron [req-2f722623-0b80-4661-8955-29293fb23871 req-86fa0c45-1bdc-4aa9-be81-babb17b7d040 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Updating instance_info_cache with network_info: [{"id": "943c04a5-7eba-46db-ab4f-93f7f84be14e", "address": "fa:16:3e:56:de:39", "network": {"id": "7635ca79-1de7-4ba3-b347-1e0b66d7eb56", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1574772814-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8e9d6e6ceac4520add365429c4eb4d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d0c6fd7-3cc9-4818-9475-8f15900394cc", "external-id": "nsx-vlan-transportzone-317", "segmentation_id": 317, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap943c04a5-7e", "ovs_interfaceid": "943c04a5-7eba-46db-ab4f-93f7f84be14e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.029770] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8bdd6669-8849-49fd-9397-746f4653b969 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.393607] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': task-1276523, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.662566} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.395401] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 4653ce96-257e-46d8-8c3d-85c03380213c/4653ce96-257e-46d8-8c3d-85c03380213c.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 812.395636] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 812.396108] env[61867]: INFO nova.compute.manager [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Took 35.99 seconds to build instance. [ 812.396895] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e1f54590-2676-48f2-a0d8-f66576061503 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.403870] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Waiting for the task: (returnval){ [ 812.403870] env[61867]: value = "task-1276524" [ 812.403870] env[61867]: _type = "Task" [ 812.403870] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.419030] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': task-1276524, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.482191] env[61867]: DEBUG oslo_concurrency.lockutils [req-2f722623-0b80-4661-8955-29293fb23871 req-86fa0c45-1bdc-4aa9-be81-babb17b7d040 service nova] Releasing lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.900829] env[61867]: DEBUG oslo_concurrency.lockutils [None req-20ae651d-e10a-43ec-a518-d283317a90a6 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "81e70c36-04ea-450c-9383-53ef069d1c46" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.414s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.919945] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': task-1276524, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066176} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.920346] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 812.921288] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a690a3b-c5c8-48ae-ad52-ff3aaede5067 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.948377] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 4653ce96-257e-46d8-8c3d-85c03380213c/4653ce96-257e-46d8-8c3d-85c03380213c.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.951543] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f37168ca-77d4-4767-9e4e-b055f224a0e4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.972096] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Waiting for the task: (returnval){ [ 812.972096] env[61867]: value = "task-1276525" [ 812.972096] env[61867]: _type = "Task" [ 812.972096] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.985072] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': task-1276525, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.174443] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da3041e-9ff0-4824-8f0a-34c54e05a6dc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.182755] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91363cda-cd89-4c3c-9cfd-d452dca8985c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.215060] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659d5a3c-e461-4250-ac6f-b41a612db851 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.223611] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6dc9d71-1222-4cef-8955-ab6b6824927a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.240291] env[61867]: DEBUG nova.compute.provider_tree [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.403679] env[61867]: DEBUG nova.compute.manager [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 813.483049] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': task-1276525, 'name': ReconfigVM_Task, 'duration_secs': 0.301392} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.483242] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 4653ce96-257e-46d8-8c3d-85c03380213c/4653ce96-257e-46d8-8c3d-85c03380213c.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.483858] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5fea0915-b833-4f23-b185-444cebc91e26 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.489939] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Waiting for the task: (returnval){ [ 813.489939] env[61867]: value = "task-1276526" [ 813.489939] env[61867]: _type = "Task" [ 813.489939] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.497684] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': task-1276526, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.746018] env[61867]: DEBUG nova.scheduler.client.report [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.929338] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.999107] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': task-1276526, 'name': Rename_Task, 'duration_secs': 0.213934} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.999500] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 813.999842] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f99cd2fb-c122-432b-a54c-9d7f4c3aa799 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.006755] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Waiting for the task: (returnval){ [ 814.006755] env[61867]: value = "task-1276527" [ 814.006755] env[61867]: _type = "Task" [ 814.006755] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.014955] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': task-1276527, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.252784] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.253443] env[61867]: DEBUG nova.compute.manager [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 814.256214] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.124s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.256432] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.258533] env[61867]: DEBUG oslo_concurrency.lockutils [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.469s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.259997] env[61867]: INFO nova.compute.claims [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.281585] env[61867]: INFO nova.scheduler.client.report [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Deleted allocations for instance c7203cea-2c2d-4a97-8ff4-3d33cd8443f7 [ 814.518292] env[61867]: DEBUG oslo_vmware.api [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': task-1276527, 'name': PowerOnVM_Task, 'duration_secs': 0.451803} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.518643] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 814.518897] env[61867]: INFO nova.compute.manager [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Took 7.11 seconds to spawn the instance on the hypervisor. [ 814.519140] env[61867]: DEBUG nova.compute.manager [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.520058] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff4d90b-f3f0-43c9-97b5-a105fc4d55d2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.765240] env[61867]: DEBUG nova.compute.utils [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 814.769572] env[61867]: DEBUG nova.compute.manager [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 814.769847] env[61867]: DEBUG nova.network.neutron [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 814.791909] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4606543e-ee26-4550-8279-4ceb1420dd30 tempest-TenantUsagesTestJSON-42359582 tempest-TenantUsagesTestJSON-42359582-project-member] Lock "c7203cea-2c2d-4a97-8ff4-3d33cd8443f7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.548s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.793101] env[61867]: DEBUG nova.compute.manager [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Stashing vm_state: active {{(pid=61867) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 814.833723] env[61867]: DEBUG nova.policy [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f772161fe6094c27ab0be84f78f8983b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '046f9c0ae5cd4a77966a57b7980326eb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 815.040596] env[61867]: INFO nova.compute.manager [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Took 34.03 seconds to build instance. [ 815.183396] env[61867]: DEBUG nova.network.neutron [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Successfully created port: 82c63e6a-f20c-460d-938c-d7af7e99764c {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 815.271064] env[61867]: DEBUG nova.compute.manager [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 815.316324] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.543389] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21ac2869-faa4-41f2-9e66-89459f1802d9 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Lock "4653ce96-257e-46d8-8c3d-85c03380213c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.776s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.557389] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32515b11-3c28-45a4-af1c-cc2783cc95f8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.565256] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87109a70-d5dd-4097-ac08-30824ce56984 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.595705] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa562685-f13b-48c5-8594-15d5e14c7d7a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.603282] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf08547-efeb-4d7c-af15-d88be0e7288a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.616870] env[61867]: DEBUG nova.compute.provider_tree [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.123031] env[61867]: DEBUG nova.scheduler.client.report [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.247835] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Acquiring lock "4653ce96-257e-46d8-8c3d-85c03380213c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.247835] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Lock "4653ce96-257e-46d8-8c3d-85c03380213c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.248920] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Acquiring lock "4653ce96-257e-46d8-8c3d-85c03380213c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.248920] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Lock "4653ce96-257e-46d8-8c3d-85c03380213c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.248920] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Lock "4653ce96-257e-46d8-8c3d-85c03380213c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.250975] env[61867]: INFO nova.compute.manager [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Terminating instance [ 816.252843] env[61867]: DEBUG nova.compute.manager [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 816.253097] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 816.253965] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70cd8b4c-ba0e-446c-8449-be7def711d0f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.262790] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 816.263132] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d6ed404-9685-4be1-8750-41a34ff02cf9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.270907] env[61867]: DEBUG oslo_vmware.api [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Waiting for the task: (returnval){ [ 816.270907] env[61867]: value = "task-1276528" [ 816.270907] env[61867]: _type = "Task" [ 816.270907] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.279620] env[61867]: DEBUG oslo_vmware.api [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': task-1276528, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.287268] env[61867]: DEBUG nova.compute.manager [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 816.321427] env[61867]: DEBUG nova.virt.hardware [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 816.321778] env[61867]: DEBUG nova.virt.hardware [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 816.321961] env[61867]: DEBUG nova.virt.hardware [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 816.322208] env[61867]: DEBUG nova.virt.hardware [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 816.322340] env[61867]: DEBUG nova.virt.hardware [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 816.322598] env[61867]: DEBUG nova.virt.hardware [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 816.322773] env[61867]: DEBUG nova.virt.hardware [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 816.323064] env[61867]: DEBUG nova.virt.hardware [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 816.323148] env[61867]: DEBUG nova.virt.hardware [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 816.323301] env[61867]: DEBUG nova.virt.hardware [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 816.323508] env[61867]: DEBUG nova.virt.hardware [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 816.324511] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b634a5-0087-4f90-bfc0-33b86c543931 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.335678] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f84866-a02c-450f-a9f1-79a4d7847a11 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.626194] env[61867]: DEBUG oslo_concurrency.lockutils [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.368s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.626791] env[61867]: DEBUG nova.compute.manager [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 816.629742] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.060s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.633181] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.635255] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.769s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.636737] env[61867]: INFO nova.compute.claims [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.661631] env[61867]: INFO nova.scheduler.client.report [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Deleted allocations for instance b1ec32c5-0642-4dca-ad43-05a5172d04d7 [ 816.785986] env[61867]: DEBUG oslo_vmware.api [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': task-1276528, 'name': PowerOffVM_Task, 'duration_secs': 0.189397} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.787122] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 816.787506] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 816.792202] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9e4c5cff-3b13-4033-bfcb-5bb931ac821a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.805309] env[61867]: DEBUG nova.compute.manager [req-e3a03a51-5683-4418-8bfa-236c25ff7c51 req-8b2c8cb3-88dc-407e-a7b7-a1dc4763f9e6 service nova] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Received event network-vif-plugged-82c63e6a-f20c-460d-938c-d7af7e99764c {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 816.805309] env[61867]: DEBUG oslo_concurrency.lockutils [req-e3a03a51-5683-4418-8bfa-236c25ff7c51 req-8b2c8cb3-88dc-407e-a7b7-a1dc4763f9e6 service nova] Acquiring lock "66518b2a-0242-438b-ba9f-d57c07a1165c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.805309] env[61867]: DEBUG oslo_concurrency.lockutils [req-e3a03a51-5683-4418-8bfa-236c25ff7c51 req-8b2c8cb3-88dc-407e-a7b7-a1dc4763f9e6 service nova] Lock "66518b2a-0242-438b-ba9f-d57c07a1165c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.805309] env[61867]: DEBUG oslo_concurrency.lockutils [req-e3a03a51-5683-4418-8bfa-236c25ff7c51 req-8b2c8cb3-88dc-407e-a7b7-a1dc4763f9e6 service nova] Lock "66518b2a-0242-438b-ba9f-d57c07a1165c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.805309] env[61867]: DEBUG nova.compute.manager [req-e3a03a51-5683-4418-8bfa-236c25ff7c51 req-8b2c8cb3-88dc-407e-a7b7-a1dc4763f9e6 service nova] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] No waiting events found dispatching network-vif-plugged-82c63e6a-f20c-460d-938c-d7af7e99764c {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 816.805790] env[61867]: WARNING nova.compute.manager [req-e3a03a51-5683-4418-8bfa-236c25ff7c51 req-8b2c8cb3-88dc-407e-a7b7-a1dc4763f9e6 service nova] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Received unexpected event network-vif-plugged-82c63e6a-f20c-460d-938c-d7af7e99764c for instance with vm_state building and task_state spawning. [ 816.857284] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 816.857601] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 816.857856] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Deleting the datastore file [datastore2] 4653ce96-257e-46d8-8c3d-85c03380213c {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 816.858151] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bc947e98-4898-4e00-a4da-c98c3d21bc71 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.864894] env[61867]: DEBUG oslo_vmware.api [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Waiting for the task: (returnval){ [ 816.864894] env[61867]: value = "task-1276530" [ 816.864894] env[61867]: _type = "Task" [ 816.864894] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.874291] env[61867]: DEBUG oslo_vmware.api [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': task-1276530, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.922492] env[61867]: DEBUG nova.network.neutron [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Successfully updated port: 82c63e6a-f20c-460d-938c-d7af7e99764c {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 817.143886] env[61867]: DEBUG nova.compute.utils [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.148484] env[61867]: DEBUG nova.compute.manager [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 817.150222] env[61867]: DEBUG nova.network.neutron [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 817.169924] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b9470a33-7bbb-45cc-82f4-3dad16ac1bfd tempest-ServerShowV247Test-1152023765 tempest-ServerShowV247Test-1152023765-project-member] Lock "b1ec32c5-0642-4dca-ad43-05a5172d04d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.306s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.235831] env[61867]: DEBUG nova.policy [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1fc8ff85bced44c3a9de8b6c71d02bd5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '250e5d2fa5cb4a8a8bd0a6fdf9e50a3d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 817.375421] env[61867]: DEBUG oslo_vmware.api [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Task: {'id': task-1276530, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.293944} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.378959] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 817.378959] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 817.378959] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 817.378959] env[61867]: INFO nova.compute.manager [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 817.378959] env[61867]: DEBUG oslo.service.loopingcall [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 817.379514] env[61867]: DEBUG nova.compute.manager [-] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 817.379514] env[61867]: DEBUG nova.network.neutron [-] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 817.428149] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Acquiring lock "refresh_cache-66518b2a-0242-438b-ba9f-d57c07a1165c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.428149] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Acquired lock "refresh_cache-66518b2a-0242-438b-ba9f-d57c07a1165c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.428149] env[61867]: DEBUG nova.network.neutron [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 817.629022] env[61867]: DEBUG nova.network.neutron [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Successfully created port: f0eb4126-40e7-4fe8-b276-192b91388aba {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 817.651982] env[61867]: DEBUG nova.compute.manager [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 817.970785] env[61867]: DEBUG nova.network.neutron [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.975769] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f994e479-e459-4852-814f-c59a7cb0fb6d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.989351] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8678a41-4733-4d4c-88d6-855eb6e74b5f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.032839] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3afce52-887f-4d07-a7ee-0e3780269f86 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.042820] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed50c4a-3614-4487-9864-0f1314c16d49 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.058416] env[61867]: DEBUG nova.compute.provider_tree [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.161049] env[61867]: DEBUG nova.network.neutron [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Updating instance_info_cache with network_info: [{"id": "82c63e6a-f20c-460d-938c-d7af7e99764c", "address": "fa:16:3e:c5:81:8f", "network": {"id": "3b4b1a5a-8952-4d1a-8ec7-2fb6ff2b4779", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-259290050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "046f9c0ae5cd4a77966a57b7980326eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82c63e6a-f2", "ovs_interfaceid": "82c63e6a-f20c-460d-938c-d7af7e99764c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.269609] env[61867]: DEBUG nova.compute.manager [req-cd9275a2-337b-49d7-bc8b-3ad033daafec req-03964021-b330-403b-ab98-8b30c7b1e2ba service nova] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Received event network-vif-deleted-5a9bc008-86f3-4487-898e-f6d09ec8f111 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 818.269609] env[61867]: INFO nova.compute.manager [req-cd9275a2-337b-49d7-bc8b-3ad033daafec req-03964021-b330-403b-ab98-8b30c7b1e2ba service nova] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Neutron deleted interface 5a9bc008-86f3-4487-898e-f6d09ec8f111; detaching it from the instance and deleting it from the info cache [ 818.269609] env[61867]: DEBUG nova.network.neutron [req-cd9275a2-337b-49d7-bc8b-3ad033daafec req-03964021-b330-403b-ab98-8b30c7b1e2ba service nova] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.562181] env[61867]: DEBUG nova.scheduler.client.report [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.614434] env[61867]: DEBUG nova.network.neutron [-] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.668040] env[61867]: DEBUG nova.compute.manager [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 818.668040] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Releasing lock "refresh_cache-66518b2a-0242-438b-ba9f-d57c07a1165c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.668104] env[61867]: DEBUG nova.compute.manager [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Instance network_info: |[{"id": "82c63e6a-f20c-460d-938c-d7af7e99764c", "address": "fa:16:3e:c5:81:8f", "network": {"id": "3b4b1a5a-8952-4d1a-8ec7-2fb6ff2b4779", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-259290050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "046f9c0ae5cd4a77966a57b7980326eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82c63e6a-f2", "ovs_interfaceid": "82c63e6a-f20c-460d-938c-d7af7e99764c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 818.668173] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:81:8f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2624812a-9f9c-461d-8b5f-79bea90c7ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '82c63e6a-f20c-460d-938c-d7af7e99764c', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 818.676726] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Creating folder: Project (046f9c0ae5cd4a77966a57b7980326eb). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.677109] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-488bc544-9ad7-4170-852b-6cb99ac44d72 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.689451] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Created folder: Project (046f9c0ae5cd4a77966a57b7980326eb) in parent group-v274258. [ 818.689629] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Creating folder: Instances. Parent ref: group-v274323. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.689870] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7357578c-084e-4d04-b085-035c5eba481e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.694709] env[61867]: DEBUG nova.virt.hardware [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 818.695057] env[61867]: DEBUG nova.virt.hardware [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 818.695252] env[61867]: DEBUG nova.virt.hardware [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.695445] env[61867]: DEBUG nova.virt.hardware [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 818.695595] env[61867]: DEBUG nova.virt.hardware [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.695963] env[61867]: DEBUG nova.virt.hardware [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 818.696214] env[61867]: DEBUG nova.virt.hardware [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 818.696384] env[61867]: DEBUG nova.virt.hardware [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 818.696559] env[61867]: DEBUG nova.virt.hardware [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 818.696780] env[61867]: DEBUG nova.virt.hardware [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 818.697918] env[61867]: DEBUG nova.virt.hardware [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.698139] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91f7d2e0-256d-4f16-8c02-cab55f19052a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.706992] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7add7388-77f9-4d3a-ba05-33e68dca7bc3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.712291] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Created folder: Instances in parent group-v274323. [ 818.713394] env[61867]: DEBUG oslo.service.loopingcall [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.713394] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 818.713529] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c8c3cfdb-35ff-4bbc-9e30-5099de4ba891 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.743757] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 818.743757] env[61867]: value = "task-1276533" [ 818.743757] env[61867]: _type = "Task" [ 818.743757] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.751718] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276533, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.773980] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-090fd239-da70-40fd-9715-e50bd3a34d0f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.783579] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ff7d02-8718-4e05-b08b-f2afe48f8bd5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.808196] env[61867]: DEBUG nova.compute.manager [req-cd9275a2-337b-49d7-bc8b-3ad033daafec req-03964021-b330-403b-ab98-8b30c7b1e2ba service nova] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Detach interface failed, port_id=5a9bc008-86f3-4487-898e-f6d09ec8f111, reason: Instance 4653ce96-257e-46d8-8c3d-85c03380213c could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 818.839193] env[61867]: DEBUG nova.compute.manager [req-ebd00923-654c-47b3-8ea6-16497f4ea1cf req-e6b3fd16-16e8-4358-a5c5-a7ecf9486146 service nova] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Received event network-changed-82c63e6a-f20c-460d-938c-d7af7e99764c {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 818.839416] env[61867]: DEBUG nova.compute.manager [req-ebd00923-654c-47b3-8ea6-16497f4ea1cf req-e6b3fd16-16e8-4358-a5c5-a7ecf9486146 service nova] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Refreshing instance network info cache due to event network-changed-82c63e6a-f20c-460d-938c-d7af7e99764c. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 818.839653] env[61867]: DEBUG oslo_concurrency.lockutils [req-ebd00923-654c-47b3-8ea6-16497f4ea1cf req-e6b3fd16-16e8-4358-a5c5-a7ecf9486146 service nova] Acquiring lock "refresh_cache-66518b2a-0242-438b-ba9f-d57c07a1165c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.839812] env[61867]: DEBUG oslo_concurrency.lockutils [req-ebd00923-654c-47b3-8ea6-16497f4ea1cf req-e6b3fd16-16e8-4358-a5c5-a7ecf9486146 service nova] Acquired lock "refresh_cache-66518b2a-0242-438b-ba9f-d57c07a1165c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.839979] env[61867]: DEBUG nova.network.neutron [req-ebd00923-654c-47b3-8ea6-16497f4ea1cf req-e6b3fd16-16e8-4358-a5c5-a7ecf9486146 service nova] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Refreshing network info cache for port 82c63e6a-f20c-460d-938c-d7af7e99764c {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 819.071153] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.436s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.071807] env[61867]: DEBUG nova.compute.manager [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 819.077026] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.067s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.077168] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.082997] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.155s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.085150] env[61867]: INFO nova.compute.claims [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 819.117408] env[61867]: INFO nova.compute.manager [-] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Took 1.74 seconds to deallocate network for instance. [ 819.133271] env[61867]: INFO nova.scheduler.client.report [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Deleted allocations for instance f0b6273f-4177-44a3-8dd1-f65faeb07539 [ 819.258826] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276533, 'name': CreateVM_Task, 'duration_secs': 0.494669} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.259102] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 819.259713] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.259886] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.261487] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 819.261487] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-050b2d0d-d4fe-4461-9aa5-4c01a8aed12a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.269452] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Waiting for the task: (returnval){ [ 819.269452] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52076325-9a93-51f3-799d-7e6dfe4410bb" [ 819.269452] env[61867]: _type = "Task" [ 819.269452] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.283337] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52076325-9a93-51f3-799d-7e6dfe4410bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.295424] env[61867]: DEBUG nova.network.neutron [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Successfully updated port: f0eb4126-40e7-4fe8-b276-192b91388aba {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 819.578488] env[61867]: DEBUG nova.compute.utils [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 819.588154] env[61867]: DEBUG nova.compute.manager [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 819.588154] env[61867]: DEBUG nova.network.neutron [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 819.629693] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.643057] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7aca759-0648-4965-a20b-5797fe39da19 tempest-InstanceActionsV221TestJSON-1533474371 tempest-InstanceActionsV221TestJSON-1533474371-project-member] Lock "f0b6273f-4177-44a3-8dd1-f65faeb07539" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.544s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.692976] env[61867]: DEBUG nova.policy [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e5f6dd2ad234411975c32648c366a7d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35aa85105b394369b17ef4ce51a49066', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 819.714039] env[61867]: DEBUG nova.network.neutron [req-ebd00923-654c-47b3-8ea6-16497f4ea1cf req-e6b3fd16-16e8-4358-a5c5-a7ecf9486146 service nova] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Updated VIF entry in instance network info cache for port 82c63e6a-f20c-460d-938c-d7af7e99764c. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 819.714039] env[61867]: DEBUG nova.network.neutron [req-ebd00923-654c-47b3-8ea6-16497f4ea1cf req-e6b3fd16-16e8-4358-a5c5-a7ecf9486146 service nova] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Updating instance_info_cache with network_info: [{"id": "82c63e6a-f20c-460d-938c-d7af7e99764c", "address": "fa:16:3e:c5:81:8f", "network": {"id": "3b4b1a5a-8952-4d1a-8ec7-2fb6ff2b4779", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-259290050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "046f9c0ae5cd4a77966a57b7980326eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82c63e6a-f2", "ovs_interfaceid": "82c63e6a-f20c-460d-938c-d7af7e99764c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.787068] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52076325-9a93-51f3-799d-7e6dfe4410bb, 'name': SearchDatastore_Task, 'duration_secs': 0.011464} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.787409] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.787649] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 819.787891] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.788052] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.789775] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 819.790156] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d02a9883-3c8e-495c-89c7-d939c770633c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.800868] env[61867]: DEBUG oslo_concurrency.lockutils [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.800868] env[61867]: DEBUG oslo_concurrency.lockutils [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.800868] env[61867]: DEBUG nova.network.neutron [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 819.801421] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 819.801690] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 819.803049] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f275f9c-a50b-406e-b9cb-0c7ab4368567 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.810658] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Waiting for the task: (returnval){ [ 819.810658] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527d1459-5a66-8b11-7ace-a3444b765e61" [ 819.810658] env[61867]: _type = "Task" [ 819.810658] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.821196] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527d1459-5a66-8b11-7ace-a3444b765e61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.083093] env[61867]: DEBUG nova.compute.manager [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 820.217245] env[61867]: DEBUG oslo_concurrency.lockutils [req-ebd00923-654c-47b3-8ea6-16497f4ea1cf req-e6b3fd16-16e8-4358-a5c5-a7ecf9486146 service nova] Releasing lock "refresh_cache-66518b2a-0242-438b-ba9f-d57c07a1165c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.322037] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527d1459-5a66-8b11-7ace-a3444b765e61, 'name': SearchDatastore_Task, 'duration_secs': 0.014222} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.325012] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34540cd6-5952-4ebe-9e0f-7cb2c4bbba5a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.331154] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Waiting for the task: (returnval){ [ 820.331154] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52244d79-3a4e-3fe0-cc88-b96272c60a91" [ 820.331154] env[61867]: _type = "Task" [ 820.331154] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.347509] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52244d79-3a4e-3fe0-cc88-b96272c60a91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.371876] env[61867]: DEBUG nova.network.neutron [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Successfully created port: d8069fac-960f-49cf-a416-036e26060a27 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 820.384793] env[61867]: DEBUG nova.network.neutron [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.452149] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943922ae-6d96-4c34-b403-21815b7d23ff {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.459932] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20f90673-14c2-44ea-8dc5-e94bccb306ea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.492524] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d08a36-6d8e-44a0-8446-b330e9248e3b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.500580] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a230514-5c98-41ce-8d07-74514abc86d5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.516925] env[61867]: DEBUG nova.compute.provider_tree [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.750936] env[61867]: DEBUG nova.network.neutron [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updating instance_info_cache with network_info: [{"id": "f0eb4126-40e7-4fe8-b276-192b91388aba", "address": "fa:16:3e:69:84:03", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0eb4126-40", "ovs_interfaceid": "f0eb4126-40e7-4fe8-b276-192b91388aba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.844669] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52244d79-3a4e-3fe0-cc88-b96272c60a91, 'name': SearchDatastore_Task, 'duration_secs': 0.022852} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.845790] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.846094] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 66518b2a-0242-438b-ba9f-d57c07a1165c/66518b2a-0242-438b-ba9f-d57c07a1165c.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 820.846827] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5af89535-a39c-48d2-847e-8759d8543a89 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.857543] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Waiting for the task: (returnval){ [ 820.857543] env[61867]: value = "task-1276534" [ 820.857543] env[61867]: _type = "Task" [ 820.857543] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.867617] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': task-1276534, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.919876] env[61867]: DEBUG nova.compute.manager [req-7c98b267-d3ec-4fd1-9f5c-b0d42762256d req-e675fac4-7c34-4944-9bff-cf51408c0a06 service nova] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Received event network-vif-plugged-f0eb4126-40e7-4fe8-b276-192b91388aba {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.921942] env[61867]: DEBUG oslo_concurrency.lockutils [req-7c98b267-d3ec-4fd1-9f5c-b0d42762256d req-e675fac4-7c34-4944-9bff-cf51408c0a06 service nova] Acquiring lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.921942] env[61867]: DEBUG oslo_concurrency.lockutils [req-7c98b267-d3ec-4fd1-9f5c-b0d42762256d req-e675fac4-7c34-4944-9bff-cf51408c0a06 service nova] Lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.921942] env[61867]: DEBUG oslo_concurrency.lockutils [req-7c98b267-d3ec-4fd1-9f5c-b0d42762256d req-e675fac4-7c34-4944-9bff-cf51408c0a06 service nova] Lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.921942] env[61867]: DEBUG nova.compute.manager [req-7c98b267-d3ec-4fd1-9f5c-b0d42762256d req-e675fac4-7c34-4944-9bff-cf51408c0a06 service nova] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] No waiting events found dispatching network-vif-plugged-f0eb4126-40e7-4fe8-b276-192b91388aba {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 820.921942] env[61867]: WARNING nova.compute.manager [req-7c98b267-d3ec-4fd1-9f5c-b0d42762256d req-e675fac4-7c34-4944-9bff-cf51408c0a06 service nova] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Received unexpected event network-vif-plugged-f0eb4126-40e7-4fe8-b276-192b91388aba for instance with vm_state building and task_state spawning. [ 820.922149] env[61867]: DEBUG nova.compute.manager [req-7c98b267-d3ec-4fd1-9f5c-b0d42762256d req-e675fac4-7c34-4944-9bff-cf51408c0a06 service nova] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Received event network-changed-f0eb4126-40e7-4fe8-b276-192b91388aba {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.922149] env[61867]: DEBUG nova.compute.manager [req-7c98b267-d3ec-4fd1-9f5c-b0d42762256d req-e675fac4-7c34-4944-9bff-cf51408c0a06 service nova] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Refreshing instance network info cache due to event network-changed-f0eb4126-40e7-4fe8-b276-192b91388aba. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 820.922149] env[61867]: DEBUG oslo_concurrency.lockutils [req-7c98b267-d3ec-4fd1-9f5c-b0d42762256d req-e675fac4-7c34-4944-9bff-cf51408c0a06 service nova] Acquiring lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.019424] env[61867]: DEBUG nova.scheduler.client.report [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 821.092443] env[61867]: DEBUG nova.compute.manager [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 821.121408] env[61867]: DEBUG nova.virt.hardware [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 821.122425] env[61867]: DEBUG nova.virt.hardware [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 821.122663] env[61867]: DEBUG nova.virt.hardware [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.123254] env[61867]: DEBUG nova.virt.hardware [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 821.123454] env[61867]: DEBUG nova.virt.hardware [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.123835] env[61867]: DEBUG nova.virt.hardware [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 821.124097] env[61867]: DEBUG nova.virt.hardware [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 821.124374] env[61867]: DEBUG nova.virt.hardware [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 821.124757] env[61867]: DEBUG nova.virt.hardware [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 821.124948] env[61867]: DEBUG nova.virt.hardware [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 821.125144] env[61867]: DEBUG nova.virt.hardware [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 821.126278] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf00c672-88e8-4b8e-ae1d-2fd181efcb7c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.135997] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b022b29d-6b94-4eb1-9585-6262701c4e70 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.253922] env[61867]: DEBUG oslo_concurrency.lockutils [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.254454] env[61867]: DEBUG nova.compute.manager [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Instance network_info: |[{"id": "f0eb4126-40e7-4fe8-b276-192b91388aba", "address": "fa:16:3e:69:84:03", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0eb4126-40", "ovs_interfaceid": "f0eb4126-40e7-4fe8-b276-192b91388aba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 821.254848] env[61867]: DEBUG oslo_concurrency.lockutils [req-7c98b267-d3ec-4fd1-9f5c-b0d42762256d req-e675fac4-7c34-4944-9bff-cf51408c0a06 service nova] Acquired lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.255009] env[61867]: DEBUG nova.network.neutron [req-7c98b267-d3ec-4fd1-9f5c-b0d42762256d req-e675fac4-7c34-4944-9bff-cf51408c0a06 service nova] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Refreshing network info cache for port f0eb4126-40e7-4fe8-b276-192b91388aba {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 821.256342] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:69:84:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '399f3826-705c-45f7-9fe0-3a08a945151a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f0eb4126-40e7-4fe8-b276-192b91388aba', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 821.270400] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Creating folder: Project (250e5d2fa5cb4a8a8bd0a6fdf9e50a3d). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 821.271533] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a5fd8b52-8699-4831-9bea-e217fc0ca82d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.289746] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Created folder: Project (250e5d2fa5cb4a8a8bd0a6fdf9e50a3d) in parent group-v274258. [ 821.289831] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Creating folder: Instances. Parent ref: group-v274326. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 821.292107] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b61563cb-2c26-442d-96db-a6b6d006b59f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.305168] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Created folder: Instances in parent group-v274326. [ 821.305577] env[61867]: DEBUG oslo.service.loopingcall [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.305728] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 821.305960] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-961ec979-ac94-4d3a-aaeb-e1743e955228 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.331821] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.331821] env[61867]: value = "task-1276537" [ 821.331821] env[61867]: _type = "Task" [ 821.331821] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.342461] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276537, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.370953] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': task-1276534, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.526876] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.527335] env[61867]: DEBUG nova.compute.manager [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 821.530850] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.658s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.534080] env[61867]: INFO nova.compute.claims [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 821.846972] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276537, 'name': CreateVM_Task, 'duration_secs': 0.410233} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.847979] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 821.849750] env[61867]: DEBUG oslo_concurrency.lockutils [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.849937] env[61867]: DEBUG oslo_concurrency.lockutils [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.850333] env[61867]: DEBUG oslo_concurrency.lockutils [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 821.850634] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf9b2ad5-d797-48a8-b51f-68e59b16488a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.855533] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 821.855533] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52912df8-119d-73f5-406a-a85fd3621319" [ 821.855533] env[61867]: _type = "Task" [ 821.855533] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.874808] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52912df8-119d-73f5-406a-a85fd3621319, 'name': SearchDatastore_Task, 'duration_secs': 0.013326} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.874967] env[61867]: DEBUG oslo_concurrency.lockutils [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.875239] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 821.876855] env[61867]: DEBUG oslo_concurrency.lockutils [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.876855] env[61867]: DEBUG oslo_concurrency.lockutils [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.876855] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 821.881081] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-47f899b4-752d-43f4-993f-b04de7123aec {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.882972] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': task-1276534, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.759549} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.883269] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 66518b2a-0242-438b-ba9f-d57c07a1165c/66518b2a-0242-438b-ba9f-d57c07a1165c.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 821.883460] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 821.885312] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18cbe8d8-24ce-4f8c-a659-be737d6da78a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.891088] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 821.891278] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 821.893218] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e4b87cb-24f0-42d8-b994-4fe85e8ff221 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.895828] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Waiting for the task: (returnval){ [ 821.895828] env[61867]: value = "task-1276538" [ 821.895828] env[61867]: _type = "Task" [ 821.895828] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.901648] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 821.901648] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52df4090-df50-b1a5-8562-961d3c50fd27" [ 821.901648] env[61867]: _type = "Task" [ 821.901648] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.911023] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': task-1276538, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.919894] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52df4090-df50-b1a5-8562-961d3c50fd27, 'name': SearchDatastore_Task, 'duration_secs': 0.012838} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.921025] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d8baa0f-6ee8-4478-9926-0b0bec1041d2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.928446] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 821.928446] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5218df6a-df1c-a3cb-858a-2cf11dc65907" [ 821.928446] env[61867]: _type = "Task" [ 821.928446] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.938960] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5218df6a-df1c-a3cb-858a-2cf11dc65907, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.041120] env[61867]: DEBUG nova.compute.utils [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.046033] env[61867]: DEBUG nova.compute.manager [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 822.050017] env[61867]: DEBUG nova.network.neutron [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 822.178731] env[61867]: DEBUG nova.policy [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e5f6dd2ad234411975c32648c366a7d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35aa85105b394369b17ef4ce51a49066', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 822.271067] env[61867]: DEBUG nova.network.neutron [req-7c98b267-d3ec-4fd1-9f5c-b0d42762256d req-e675fac4-7c34-4944-9bff-cf51408c0a06 service nova] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updated VIF entry in instance network info cache for port f0eb4126-40e7-4fe8-b276-192b91388aba. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 822.271489] env[61867]: DEBUG nova.network.neutron [req-7c98b267-d3ec-4fd1-9f5c-b0d42762256d req-e675fac4-7c34-4944-9bff-cf51408c0a06 service nova] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updating instance_info_cache with network_info: [{"id": "f0eb4126-40e7-4fe8-b276-192b91388aba", "address": "fa:16:3e:69:84:03", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0eb4126-40", "ovs_interfaceid": "f0eb4126-40e7-4fe8-b276-192b91388aba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.322405] env[61867]: DEBUG oslo_vmware.rw_handles [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529c83e1-d237-3f7b-55b1-69622459ec92/disk-0.vmdk. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 822.323914] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f47565-8e7f-4770-9843-8efe81271767 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.330984] env[61867]: DEBUG oslo_vmware.rw_handles [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529c83e1-d237-3f7b-55b1-69622459ec92/disk-0.vmdk is in state: ready. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 822.331208] env[61867]: ERROR oslo_vmware.rw_handles [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529c83e1-d237-3f7b-55b1-69622459ec92/disk-0.vmdk due to incomplete transfer. [ 822.331496] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-45285e5b-f82d-4c20-b351-739ca3c1a6e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.339450] env[61867]: DEBUG oslo_vmware.rw_handles [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529c83e1-d237-3f7b-55b1-69622459ec92/disk-0.vmdk. {{(pid=61867) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 822.339827] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Uploaded image 420f5356-3ba1-4114-af3a-5be2a65b82cb to the Glance image server {{(pid=61867) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 822.341933] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Destroying the VM {{(pid=61867) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 822.342848] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-315078c2-a198-464c-8ae9-141d87b38449 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.348705] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 822.348705] env[61867]: value = "task-1276539" [ 822.348705] env[61867]: _type = "Task" [ 822.348705] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.358417] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276539, 'name': Destroy_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.407311] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': task-1276538, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078404} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.407806] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 822.408739] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd922ea5-35bc-43cf-a03f-76c657badf50 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.437205] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 66518b2a-0242-438b-ba9f-d57c07a1165c/66518b2a-0242-438b-ba9f-d57c07a1165c.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 822.437616] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9dc0681f-b5d3-4367-a4fe-6b3c0d76aaa3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.467077] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5218df6a-df1c-a3cb-858a-2cf11dc65907, 'name': SearchDatastore_Task, 'duration_secs': 0.012857} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.468564] env[61867]: DEBUG oslo_concurrency.lockutils [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.469107] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] abb41c0c-6d0d-4147-a4af-554ab7d9e921/abb41c0c-6d0d-4147-a4af-554ab7d9e921.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 822.469514] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Waiting for the task: (returnval){ [ 822.469514] env[61867]: value = "task-1276540" [ 822.469514] env[61867]: _type = "Task" [ 822.469514] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.469766] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3720ba29-92ac-4cd4-b528-e194a809a6c0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.485463] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': task-1276540, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.485463] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 822.485463] env[61867]: value = "task-1276541" [ 822.485463] env[61867]: _type = "Task" [ 822.485463] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.494439] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276541, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.529082] env[61867]: DEBUG nova.network.neutron [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Successfully updated port: d8069fac-960f-49cf-a416-036e26060a27 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 822.546873] env[61867]: DEBUG nova.compute.manager [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 822.776886] env[61867]: DEBUG oslo_concurrency.lockutils [req-7c98b267-d3ec-4fd1-9f5c-b0d42762256d req-e675fac4-7c34-4944-9bff-cf51408c0a06 service nova] Releasing lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.825470] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-639c9c7b-76e4-40f6-af6a-4e2075826c56 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.834509] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d61c4d-d788-4202-a684-362c1d91f854 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.881145] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0138ea3-1d93-4f5c-b6bb-da8b5c911b4a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.883661] env[61867]: DEBUG nova.network.neutron [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Successfully created port: cdf39635-188f-4031-ae51-c7d98a8e8365 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 822.892542] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276539, 'name': Destroy_Task, 'duration_secs': 0.356364} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.894783] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Destroyed the VM [ 822.895220] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Deleting Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 822.895566] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-bc0e04df-6a08-457a-a606-4d726eea0a4f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.898472] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aba4c5d-3147-4e67-a030-83aaac0397d8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.913409] env[61867]: DEBUG nova.compute.provider_tree [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.916926] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 822.916926] env[61867]: value = "task-1276542" [ 822.916926] env[61867]: _type = "Task" [ 822.916926] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.929940] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276542, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.954104] env[61867]: DEBUG nova.compute.manager [req-9be8a5e6-f8f3-4b8e-bb06-cf8963526dbf req-0a42ef99-712b-4995-b8d9-63f0c79652d8 service nova] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Received event network-vif-plugged-d8069fac-960f-49cf-a416-036e26060a27 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.954194] env[61867]: DEBUG oslo_concurrency.lockutils [req-9be8a5e6-f8f3-4b8e-bb06-cf8963526dbf req-0a42ef99-712b-4995-b8d9-63f0c79652d8 service nova] Acquiring lock "6e41989e-b8fa-4009-af1e-1ce859b329a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.954558] env[61867]: DEBUG oslo_concurrency.lockutils [req-9be8a5e6-f8f3-4b8e-bb06-cf8963526dbf req-0a42ef99-712b-4995-b8d9-63f0c79652d8 service nova] Lock "6e41989e-b8fa-4009-af1e-1ce859b329a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.954715] env[61867]: DEBUG oslo_concurrency.lockutils [req-9be8a5e6-f8f3-4b8e-bb06-cf8963526dbf req-0a42ef99-712b-4995-b8d9-63f0c79652d8 service nova] Lock "6e41989e-b8fa-4009-af1e-1ce859b329a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.954889] env[61867]: DEBUG nova.compute.manager [req-9be8a5e6-f8f3-4b8e-bb06-cf8963526dbf req-0a42ef99-712b-4995-b8d9-63f0c79652d8 service nova] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] No waiting events found dispatching network-vif-plugged-d8069fac-960f-49cf-a416-036e26060a27 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 822.955079] env[61867]: WARNING nova.compute.manager [req-9be8a5e6-f8f3-4b8e-bb06-cf8963526dbf req-0a42ef99-712b-4995-b8d9-63f0c79652d8 service nova] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Received unexpected event network-vif-plugged-d8069fac-960f-49cf-a416-036e26060a27 for instance with vm_state building and task_state spawning. [ 822.955239] env[61867]: DEBUG nova.compute.manager [req-9be8a5e6-f8f3-4b8e-bb06-cf8963526dbf req-0a42ef99-712b-4995-b8d9-63f0c79652d8 service nova] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Received event network-changed-d8069fac-960f-49cf-a416-036e26060a27 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.955703] env[61867]: DEBUG nova.compute.manager [req-9be8a5e6-f8f3-4b8e-bb06-cf8963526dbf req-0a42ef99-712b-4995-b8d9-63f0c79652d8 service nova] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Refreshing instance network info cache due to event network-changed-d8069fac-960f-49cf-a416-036e26060a27. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 822.955703] env[61867]: DEBUG oslo_concurrency.lockutils [req-9be8a5e6-f8f3-4b8e-bb06-cf8963526dbf req-0a42ef99-712b-4995-b8d9-63f0c79652d8 service nova] Acquiring lock "refresh_cache-6e41989e-b8fa-4009-af1e-1ce859b329a1" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.955907] env[61867]: DEBUG oslo_concurrency.lockutils [req-9be8a5e6-f8f3-4b8e-bb06-cf8963526dbf req-0a42ef99-712b-4995-b8d9-63f0c79652d8 service nova] Acquired lock "refresh_cache-6e41989e-b8fa-4009-af1e-1ce859b329a1" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.956326] env[61867]: DEBUG nova.network.neutron [req-9be8a5e6-f8f3-4b8e-bb06-cf8963526dbf req-0a42ef99-712b-4995-b8d9-63f0c79652d8 service nova] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Refreshing network info cache for port d8069fac-960f-49cf-a416-036e26060a27 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 822.989733] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': task-1276540, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.000357] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276541, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.035756] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "refresh_cache-6e41989e-b8fa-4009-af1e-1ce859b329a1" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.419233] env[61867]: DEBUG nova.scheduler.client.report [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.436922] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276542, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.493029] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': task-1276540, 'name': ReconfigVM_Task, 'duration_secs': 0.670521} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.493668] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 66518b2a-0242-438b-ba9f-d57c07a1165c/66518b2a-0242-438b-ba9f-d57c07a1165c.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 823.495568] env[61867]: DEBUG nova.network.neutron [req-9be8a5e6-f8f3-4b8e-bb06-cf8963526dbf req-0a42ef99-712b-4995-b8d9-63f0c79652d8 service nova] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.497462] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2cb8c999-64b9-4620-8b4c-f0c28e170767 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.505665] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276541, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.548423} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.507031] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] abb41c0c-6d0d-4147-a4af-554ab7d9e921/abb41c0c-6d0d-4147-a4af-554ab7d9e921.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 823.507268] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 823.507660] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Waiting for the task: (returnval){ [ 823.507660] env[61867]: value = "task-1276543" [ 823.507660] env[61867]: _type = "Task" [ 823.507660] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.508042] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7eaa0582-80cb-4643-994d-a24007198114 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.518812] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': task-1276543, 'name': Rename_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.519614] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 823.519614] env[61867]: value = "task-1276544" [ 823.519614] env[61867]: _type = "Task" [ 823.519614] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.528245] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276544, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.562081] env[61867]: DEBUG nova.compute.manager [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 823.583546] env[61867]: DEBUG nova.network.neutron [req-9be8a5e6-f8f3-4b8e-bb06-cf8963526dbf req-0a42ef99-712b-4995-b8d9-63f0c79652d8 service nova] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.587329] env[61867]: DEBUG nova.virt.hardware [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 823.587574] env[61867]: DEBUG nova.virt.hardware [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 823.587737] env[61867]: DEBUG nova.virt.hardware [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 823.587920] env[61867]: DEBUG nova.virt.hardware [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 823.588082] env[61867]: DEBUG nova.virt.hardware [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 823.588239] env[61867]: DEBUG nova.virt.hardware [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 823.588488] env[61867]: DEBUG nova.virt.hardware [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 823.588668] env[61867]: DEBUG nova.virt.hardware [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 823.588838] env[61867]: DEBUG nova.virt.hardware [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 823.589009] env[61867]: DEBUG nova.virt.hardware [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 823.589191] env[61867]: DEBUG nova.virt.hardware [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 823.590068] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f028815-58c7-43ba-9e50-3cfbfcd09ae7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.598501] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2915063-4501-482a-b4b3-4822e8a8e7e1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.939322] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.406s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.939322] env[61867]: DEBUG nova.compute.manager [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 823.944458] env[61867]: DEBUG oslo_vmware.api [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276542, 'name': RemoveSnapshot_Task, 'duration_secs': 0.785367} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.945562] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.647s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.947454] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.950365] env[61867]: DEBUG oslo_concurrency.lockutils [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.846s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.950581] env[61867]: DEBUG oslo_concurrency.lockutils [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.952692] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.823s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.955446] env[61867]: INFO nova.compute.claims [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.962041] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Deleted Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 823.962041] env[61867]: INFO nova.compute.manager [None req-434123fb-ffbc-4f5c-9594-2d92a049f50d tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Took 17.85 seconds to snapshot the instance on the hypervisor. [ 824.025917] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': task-1276543, 'name': Rename_Task, 'duration_secs': 0.249125} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.030121] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 824.031626] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f94a7b0-6424-4d1d-8a82-9d0284654c02 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.036565] env[61867]: INFO nova.scheduler.client.report [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleted allocations for instance 3022ea79-7c27-40a4-80d6-1a86b354917a [ 824.044647] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276544, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096364} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.047025] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 824.047634] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Waiting for the task: (returnval){ [ 824.047634] env[61867]: value = "task-1276545" [ 824.047634] env[61867]: _type = "Task" [ 824.047634] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.048366] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ff95dd-735b-4219-b1b7-173355555fd8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.076396] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] abb41c0c-6d0d-4147-a4af-554ab7d9e921/abb41c0c-6d0d-4147-a4af-554ab7d9e921.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 824.077893] env[61867]: INFO nova.scheduler.client.report [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Deleted allocations for instance 905ee661-d5bd-4121-9bd0-fe68bdabd2fd [ 824.082223] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb42abe1-e612-4421-8d27-3c99fad3edd2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.096734] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': task-1276545, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.099883] env[61867]: DEBUG oslo_concurrency.lockutils [req-9be8a5e6-f8f3-4b8e-bb06-cf8963526dbf req-0a42ef99-712b-4995-b8d9-63f0c79652d8 service nova] Releasing lock "refresh_cache-6e41989e-b8fa-4009-af1e-1ce859b329a1" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.100430] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired lock "refresh_cache-6e41989e-b8fa-4009-af1e-1ce859b329a1" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.100591] env[61867]: DEBUG nova.network.neutron [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 824.106499] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 824.106499] env[61867]: value = "task-1276546" [ 824.106499] env[61867]: _type = "Task" [ 824.106499] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.123879] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276546, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.468211] env[61867]: DEBUG nova.compute.utils [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 824.474329] env[61867]: DEBUG nova.compute.manager [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 824.474607] env[61867]: DEBUG nova.network.neutron [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 824.533802] env[61867]: DEBUG nova.policy [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1a29e34abfc4ce7873f7bf7d96eb474', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd177558d263c4c8295ef9406bd4aee99', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 824.544627] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2054425f-cc41-4609-a597-e08ad4713d59 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "3022ea79-7c27-40a4-80d6-1a86b354917a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.273s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.561208] env[61867]: DEBUG oslo_vmware.api [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': task-1276545, 'name': PowerOnVM_Task, 'duration_secs': 0.497318} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.561709] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 824.561779] env[61867]: INFO nova.compute.manager [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Took 8.27 seconds to spawn the instance on the hypervisor. [ 824.561921] env[61867]: DEBUG nova.compute.manager [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 824.563299] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33cf245-e1f2-4d57-bc53-dddccbed9e34 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.586754] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.587012] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.605351] env[61867]: DEBUG oslo_concurrency.lockutils [None req-949c8b0e-c5fd-4187-b987-70befaf05e1d tempest-SecurityGroupsTestJSON-1602396988 tempest-SecurityGroupsTestJSON-1602396988-project-member] Lock "905ee661-d5bd-4121-9bd0-fe68bdabd2fd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.546s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.620059] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276546, 'name': ReconfigVM_Task, 'duration_secs': 0.273719} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.620059] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Reconfigured VM instance instance-00000042 to attach disk [datastore1] abb41c0c-6d0d-4147-a4af-554ab7d9e921/abb41c0c-6d0d-4147-a4af-554ab7d9e921.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 824.620059] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1a0ae419-4d35-43db-bbaa-a54f1c3a290e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.627416] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 824.627416] env[61867]: value = "task-1276547" [ 824.627416] env[61867]: _type = "Task" [ 824.627416] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.638337] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276547, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.641120] env[61867]: DEBUG nova.network.neutron [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.906666] env[61867]: DEBUG nova.network.neutron [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Updating instance_info_cache with network_info: [{"id": "d8069fac-960f-49cf-a416-036e26060a27", "address": "fa:16:3e:ca:61:fe", "network": {"id": "ab6ec5b9-8193-43c9-a3ce-d63f31e5a001", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1573384141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35aa85105b394369b17ef4ce51a49066", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8069fac-96", "ovs_interfaceid": "d8069fac-960f-49cf-a416-036e26060a27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.973053] env[61867]: DEBUG nova.compute.manager [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 824.999191] env[61867]: DEBUG nova.network.neutron [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Successfully created port: eaffa9e8-2616-4b49-9668-c8cc7ef884c0 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 825.085528] env[61867]: INFO nova.compute.manager [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Took 43.25 seconds to build instance. [ 825.092324] env[61867]: DEBUG nova.compute.manager [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 825.138937] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276547, 'name': Rename_Task, 'duration_secs': 0.287046} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.139247] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 825.139503] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8787edab-6bed-426c-8ed8-8e30a3f9a517 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.147815] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 825.147815] env[61867]: value = "task-1276548" [ 825.147815] env[61867]: _type = "Task" [ 825.147815] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.158711] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276548, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.245844] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d03534-8450-45fe-83b2-350273fed6bc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.255755] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-badfa399-ec9c-430c-a93d-d6a9a3accfc5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.293692] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f5f4d0f-3a97-4b41-8421-1e63c47e9c91 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.301906] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec1db4b-e86d-4c14-88d4-1594163ab10c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.316813] env[61867]: DEBUG nova.compute.provider_tree [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.413015] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Releasing lock "refresh_cache-6e41989e-b8fa-4009-af1e-1ce859b329a1" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.413015] env[61867]: DEBUG nova.compute.manager [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Instance network_info: |[{"id": "d8069fac-960f-49cf-a416-036e26060a27", "address": "fa:16:3e:ca:61:fe", "network": {"id": "ab6ec5b9-8193-43c9-a3ce-d63f31e5a001", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1573384141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35aa85105b394369b17ef4ce51a49066", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8069fac-96", "ovs_interfaceid": "d8069fac-960f-49cf-a416-036e26060a27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 825.413278] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:61:fe', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ead20342-9afa-435e-a22b-b4a903457712', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd8069fac-960f-49cf-a416-036e26060a27', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 825.419967] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Creating folder: Project (35aa85105b394369b17ef4ce51a49066). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.420480] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-15a5a9bc-341b-4977-9b1c-b7d6e61db0c6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.432843] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Created folder: Project (35aa85105b394369b17ef4ce51a49066) in parent group-v274258. [ 825.432843] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Creating folder: Instances. Parent ref: group-v274329. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.432843] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53c28cd4-f464-4710-a950-14503293041b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.444395] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Created folder: Instances in parent group-v274329. [ 825.444395] env[61867]: DEBUG oslo.service.loopingcall [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.444395] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 825.444395] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e39382af-4d95-49ef-a880-0c86fd544b85 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.467040] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 825.467040] env[61867]: value = "task-1276551" [ 825.467040] env[61867]: _type = "Task" [ 825.467040] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.478186] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276551, 'name': CreateVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.573898] env[61867]: DEBUG nova.network.neutron [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Successfully updated port: cdf39635-188f-4031-ae51-c7d98a8e8365 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 825.590742] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0d40a9cf-ebf1-43a8-bf98-97424eb2b33b tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Lock "66518b2a-0242-438b-ba9f-d57c07a1165c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.381s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.616674] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.660553] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276548, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.692775] env[61867]: DEBUG nova.compute.manager [req-f7b74faf-8b4e-4527-a826-7d2670bf97df req-e6a1dea7-f016-4278-9353-0a6e4c917bb8 service nova] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Received event network-vif-plugged-cdf39635-188f-4031-ae51-c7d98a8e8365 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.692946] env[61867]: DEBUG oslo_concurrency.lockutils [req-f7b74faf-8b4e-4527-a826-7d2670bf97df req-e6a1dea7-f016-4278-9353-0a6e4c917bb8 service nova] Acquiring lock "8a83f4f2-58eb-473e-9b1e-32ce633554f9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.693163] env[61867]: DEBUG oslo_concurrency.lockutils [req-f7b74faf-8b4e-4527-a826-7d2670bf97df req-e6a1dea7-f016-4278-9353-0a6e4c917bb8 service nova] Lock "8a83f4f2-58eb-473e-9b1e-32ce633554f9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.693330] env[61867]: DEBUG oslo_concurrency.lockutils [req-f7b74faf-8b4e-4527-a826-7d2670bf97df req-e6a1dea7-f016-4278-9353-0a6e4c917bb8 service nova] Lock "8a83f4f2-58eb-473e-9b1e-32ce633554f9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.693693] env[61867]: DEBUG nova.compute.manager [req-f7b74faf-8b4e-4527-a826-7d2670bf97df req-e6a1dea7-f016-4278-9353-0a6e4c917bb8 service nova] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] No waiting events found dispatching network-vif-plugged-cdf39635-188f-4031-ae51-c7d98a8e8365 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 825.694008] env[61867]: WARNING nova.compute.manager [req-f7b74faf-8b4e-4527-a826-7d2670bf97df req-e6a1dea7-f016-4278-9353-0a6e4c917bb8 service nova] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Received unexpected event network-vif-plugged-cdf39635-188f-4031-ae51-c7d98a8e8365 for instance with vm_state building and task_state spawning. [ 825.820150] env[61867]: DEBUG nova.scheduler.client.report [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.870418] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.870686] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.877374] env[61867]: DEBUG oslo_concurrency.lockutils [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "6d2dab88-4165-4952-8019-2eaf3b863115" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.877779] env[61867]: DEBUG oslo_concurrency.lockutils [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "6d2dab88-4165-4952-8019-2eaf3b863115" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.979934] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276551, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.988028] env[61867]: DEBUG nova.compute.manager [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 826.021250] env[61867]: DEBUG nova.virt.hardware [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 826.021991] env[61867]: DEBUG nova.virt.hardware [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 826.021991] env[61867]: DEBUG nova.virt.hardware [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 826.022108] env[61867]: DEBUG nova.virt.hardware [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 826.022335] env[61867]: DEBUG nova.virt.hardware [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 826.022571] env[61867]: DEBUG nova.virt.hardware [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 826.022885] env[61867]: DEBUG nova.virt.hardware [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 826.023135] env[61867]: DEBUG nova.virt.hardware [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 826.023394] env[61867]: DEBUG nova.virt.hardware [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 826.023656] env[61867]: DEBUG nova.virt.hardware [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 826.024115] env[61867]: DEBUG nova.virt.hardware [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 826.025408] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4005303a-c85e-48b3-946d-9d4ff6200f96 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.035457] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56c8999-a0f8-4916-9367-65bec813ea88 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.083557] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "refresh_cache-8a83f4f2-58eb-473e-9b1e-32ce633554f9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.083760] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired lock "refresh_cache-8a83f4f2-58eb-473e-9b1e-32ce633554f9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.083925] env[61867]: DEBUG nova.network.neutron [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 826.162375] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276548, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.326058] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.327474] env[61867]: DEBUG nova.compute.manager [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 826.333694] env[61867]: DEBUG oslo_concurrency.lockutils [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.792s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.336227] env[61867]: INFO nova.compute.claims [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.378972] env[61867]: DEBUG nova.compute.manager [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 826.483591] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276551, 'name': CreateVM_Task, 'duration_secs': 0.548219} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.484222] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 826.484505] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.484683] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.485019] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 826.485280] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a5d4fae-47cd-4b79-a813-1750f3f40d20 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.493634] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 826.493634] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525c826e-6bd0-6154-47eb-9810cf4a9782" [ 826.493634] env[61867]: _type = "Task" [ 826.493634] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.504614] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525c826e-6bd0-6154-47eb-9810cf4a9782, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.662459] env[61867]: DEBUG oslo_vmware.api [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276548, 'name': PowerOnVM_Task, 'duration_secs': 1.181727} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.662846] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 826.662951] env[61867]: INFO nova.compute.manager [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Took 8.00 seconds to spawn the instance on the hypervisor. [ 826.664705] env[61867]: DEBUG nova.compute.manager [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 826.664705] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d160a1f-061a-4ec7-b2e0-59abb61c506f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.824109] env[61867]: DEBUG nova.network.neutron [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.833722] env[61867]: DEBUG nova.network.neutron [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Successfully updated port: eaffa9e8-2616-4b49-9668-c8cc7ef884c0 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 826.838325] env[61867]: DEBUG nova.compute.utils [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.838325] env[61867]: DEBUG nova.compute.manager [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 826.838325] env[61867]: DEBUG nova.network.neutron [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 826.889343] env[61867]: DEBUG nova.policy [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12ed008ec3204102b2b08e61ed24f418', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '47c5b5b783ea445eb1c02fb728b1fff2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 826.914905] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.007208] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525c826e-6bd0-6154-47eb-9810cf4a9782, 'name': SearchDatastore_Task, 'duration_secs': 0.030646} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.007736] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.008149] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 827.008427] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.008667] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.009028] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 827.009343] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-71024344-6bfe-4775-b6cb-924a69ff11e2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.021761] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 827.021977] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 827.023114] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-320bd00a-9fa2-4abd-9a55-58d1aafc2a3f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.029398] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 827.029398] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52aae3e2-1124-1391-8192-fa3024952f44" [ 827.029398] env[61867]: _type = "Task" [ 827.029398] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.037692] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52aae3e2-1124-1391-8192-fa3024952f44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.096546] env[61867]: DEBUG nova.network.neutron [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Updating instance_info_cache with network_info: [{"id": "cdf39635-188f-4031-ae51-c7d98a8e8365", "address": "fa:16:3e:01:31:7c", "network": {"id": "ab6ec5b9-8193-43c9-a3ce-d63f31e5a001", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1573384141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35aa85105b394369b17ef4ce51a49066", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdf39635-18", "ovs_interfaceid": "cdf39635-188f-4031-ae51-c7d98a8e8365", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.187419] env[61867]: INFO nova.compute.manager [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Took 43.42 seconds to build instance. [ 827.196251] env[61867]: DEBUG nova.network.neutron [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Successfully created port: 104495f6-b976-4f53-9959-d5193f833ae8 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 827.205015] env[61867]: DEBUG nova.compute.manager [req-1e114771-e5e2-4199-9eec-40eabb12fef7 req-778302c8-55ed-492f-9b69-4929ac6fabeb service nova] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Received event network-vif-plugged-eaffa9e8-2616-4b49-9668-c8cc7ef884c0 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.205267] env[61867]: DEBUG oslo_concurrency.lockutils [req-1e114771-e5e2-4199-9eec-40eabb12fef7 req-778302c8-55ed-492f-9b69-4929ac6fabeb service nova] Acquiring lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.206849] env[61867]: DEBUG oslo_concurrency.lockutils [req-1e114771-e5e2-4199-9eec-40eabb12fef7 req-778302c8-55ed-492f-9b69-4929ac6fabeb service nova] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.208835] env[61867]: DEBUG oslo_concurrency.lockutils [req-1e114771-e5e2-4199-9eec-40eabb12fef7 req-778302c8-55ed-492f-9b69-4929ac6fabeb service nova] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.208835] env[61867]: DEBUG nova.compute.manager [req-1e114771-e5e2-4199-9eec-40eabb12fef7 req-778302c8-55ed-492f-9b69-4929ac6fabeb service nova] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] No waiting events found dispatching network-vif-plugged-eaffa9e8-2616-4b49-9668-c8cc7ef884c0 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 827.208835] env[61867]: WARNING nova.compute.manager [req-1e114771-e5e2-4199-9eec-40eabb12fef7 req-778302c8-55ed-492f-9b69-4929ac6fabeb service nova] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Received unexpected event network-vif-plugged-eaffa9e8-2616-4b49-9668-c8cc7ef884c0 for instance with vm_state building and task_state spawning. [ 827.342285] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "refresh_cache-6dda389d-a8c5-4e0e-87a5-4065e24c034e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.342285] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquired lock "refresh_cache-6dda389d-a8c5-4e0e-87a5-4065e24c034e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.342285] env[61867]: DEBUG nova.network.neutron [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 827.342285] env[61867]: DEBUG nova.compute.manager [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 827.542295] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52aae3e2-1124-1391-8192-fa3024952f44, 'name': SearchDatastore_Task, 'duration_secs': 0.01206} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.545633] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b964419-0516-4089-a898-af021d4a27ef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.552908] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 827.552908] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52df5dcb-a5f1-e886-659c-3f7490a8be57" [ 827.552908] env[61867]: _type = "Task" [ 827.552908] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.559233] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52df5dcb-a5f1-e886-659c-3f7490a8be57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.599503] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Releasing lock "refresh_cache-8a83f4f2-58eb-473e-9b1e-32ce633554f9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.600249] env[61867]: DEBUG nova.compute.manager [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Instance network_info: |[{"id": "cdf39635-188f-4031-ae51-c7d98a8e8365", "address": "fa:16:3e:01:31:7c", "network": {"id": "ab6ec5b9-8193-43c9-a3ce-d63f31e5a001", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1573384141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35aa85105b394369b17ef4ce51a49066", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdf39635-18", "ovs_interfaceid": "cdf39635-188f-4031-ae51-c7d98a8e8365", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 827.600510] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:31:7c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ead20342-9afa-435e-a22b-b4a903457712', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cdf39635-188f-4031-ae51-c7d98a8e8365', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 827.608212] env[61867]: DEBUG oslo.service.loopingcall [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.610773] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 827.614025] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-48216081-f60a-41e0-aed3-61a98d5f4945 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.641501] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 827.641501] env[61867]: value = "task-1276552" [ 827.641501] env[61867]: _type = "Task" [ 827.641501] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.649839] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276552, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.666118] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-700f4652-9a25-4aca-ac16-37959c9aa67a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.672892] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a22cc7f-68fa-4fff-8835-f5cab59592f3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.707458] env[61867]: DEBUG oslo_concurrency.lockutils [None req-60517434-14c7-4b35-9c48-b55a068dc01e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.584s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.709781] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f1328e5-d4a2-42ec-b308-459c984bbfea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.720340] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75ab5bcf-8a3f-4508-acfc-cf03e28e5364 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.739332] env[61867]: DEBUG nova.compute.provider_tree [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.742151] env[61867]: DEBUG nova.compute.manager [req-715d13e6-61a3-45e3-9994-bbecf277537a req-18e41dc3-db05-4272-b6be-c914753cd3c0 service nova] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Received event network-changed-cdf39635-188f-4031-ae51-c7d98a8e8365 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.742295] env[61867]: DEBUG nova.compute.manager [req-715d13e6-61a3-45e3-9994-bbecf277537a req-18e41dc3-db05-4272-b6be-c914753cd3c0 service nova] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Refreshing instance network info cache due to event network-changed-cdf39635-188f-4031-ae51-c7d98a8e8365. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 827.742570] env[61867]: DEBUG oslo_concurrency.lockutils [req-715d13e6-61a3-45e3-9994-bbecf277537a req-18e41dc3-db05-4272-b6be-c914753cd3c0 service nova] Acquiring lock "refresh_cache-8a83f4f2-58eb-473e-9b1e-32ce633554f9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.742782] env[61867]: DEBUG oslo_concurrency.lockutils [req-715d13e6-61a3-45e3-9994-bbecf277537a req-18e41dc3-db05-4272-b6be-c914753cd3c0 service nova] Acquired lock "refresh_cache-8a83f4f2-58eb-473e-9b1e-32ce633554f9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.743008] env[61867]: DEBUG nova.network.neutron [req-715d13e6-61a3-45e3-9994-bbecf277537a req-18e41dc3-db05-4272-b6be-c914753cd3c0 service nova] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Refreshing network info cache for port cdf39635-188f-4031-ae51-c7d98a8e8365 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 827.921074] env[61867]: DEBUG nova.network.neutron [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.066098] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52df5dcb-a5f1-e886-659c-3f7490a8be57, 'name': SearchDatastore_Task, 'duration_secs': 0.033697} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.066431] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.066653] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 6e41989e-b8fa-4009-af1e-1ce859b329a1/6e41989e-b8fa-4009-af1e-1ce859b329a1.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 828.067024] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49ad1c11-72a1-46fb-a388-d515ac8971a9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.073725] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 828.073725] env[61867]: value = "task-1276553" [ 828.073725] env[61867]: _type = "Task" [ 828.073725] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.081529] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276553, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.151571] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276552, 'name': CreateVM_Task, 'duration_secs': 0.349464} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.151755] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 828.152384] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.152604] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.152909] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 828.153184] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6680b71b-15b8-48e6-a84e-35cd528eff4e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.157895] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 828.157895] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523c9865-cc0e-614f-3418-292321acb531" [ 828.157895] env[61867]: _type = "Task" [ 828.157895] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.165442] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523c9865-cc0e-614f-3418-292321acb531, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.213791] env[61867]: DEBUG nova.compute.manager [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 828.246968] env[61867]: DEBUG nova.scheduler.client.report [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.353345] env[61867]: DEBUG nova.compute.manager [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 828.397311] env[61867]: DEBUG nova.virt.hardware [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.397900] env[61867]: DEBUG nova.virt.hardware [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.397900] env[61867]: DEBUG nova.virt.hardware [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.397900] env[61867]: DEBUG nova.virt.hardware [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.398110] env[61867]: DEBUG nova.virt.hardware [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.398217] env[61867]: DEBUG nova.virt.hardware [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.398455] env[61867]: DEBUG nova.virt.hardware [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.398621] env[61867]: DEBUG nova.virt.hardware [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.399909] env[61867]: DEBUG nova.virt.hardware [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.399909] env[61867]: DEBUG nova.virt.hardware [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.399909] env[61867]: DEBUG nova.virt.hardware [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.400238] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6304f9e2-d0bc-4b6b-a68a-77c231c9b531 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.408784] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb2b061-cf56-4aa6-aba8-5d478edf04ba {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.502807] env[61867]: DEBUG nova.network.neutron [req-715d13e6-61a3-45e3-9994-bbecf277537a req-18e41dc3-db05-4272-b6be-c914753cd3c0 service nova] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Updated VIF entry in instance network info cache for port cdf39635-188f-4031-ae51-c7d98a8e8365. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 828.503722] env[61867]: DEBUG nova.network.neutron [req-715d13e6-61a3-45e3-9994-bbecf277537a req-18e41dc3-db05-4272-b6be-c914753cd3c0 service nova] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Updating instance_info_cache with network_info: [{"id": "cdf39635-188f-4031-ae51-c7d98a8e8365", "address": "fa:16:3e:01:31:7c", "network": {"id": "ab6ec5b9-8193-43c9-a3ce-d63f31e5a001", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1573384141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35aa85105b394369b17ef4ce51a49066", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdf39635-18", "ovs_interfaceid": "cdf39635-188f-4031-ae51-c7d98a8e8365", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.540366] env[61867]: DEBUG nova.network.neutron [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Updating instance_info_cache with network_info: [{"id": "eaffa9e8-2616-4b49-9668-c8cc7ef884c0", "address": "fa:16:3e:ec:73:fc", "network": {"id": "ee84973a-7c75-4c60-8bb4-8da65e165668", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-732335052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d177558d263c4c8295ef9406bd4aee99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fa01fe1a-83b6-4c10-af75-00ddb17f9bbf", "external-id": "nsx-vlan-transportzone-431", "segmentation_id": 431, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeaffa9e8-26", "ovs_interfaceid": "eaffa9e8-2616-4b49-9668-c8cc7ef884c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.586436] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276553, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.674905] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523c9865-cc0e-614f-3418-292321acb531, 'name': SearchDatastore_Task, 'duration_secs': 0.010718} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.674905] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.675295] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 828.675657] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.675922] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.676983] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 828.676983] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-123a2068-715f-47b6-a656-80109ce2daa2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.688675] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 828.691021] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 828.691021] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c36125a3-a475-43c4-ae40-33fd0da42822 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.697313] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 828.697313] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5248c0f4-5c05-0b62-14e3-2643578b4f02" [ 828.697313] env[61867]: _type = "Task" [ 828.697313] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.707103] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5248c0f4-5c05-0b62-14e3-2643578b4f02, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.739923] env[61867]: DEBUG oslo_concurrency.lockutils [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.760264] env[61867]: DEBUG oslo_concurrency.lockutils [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.424s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.760264] env[61867]: DEBUG nova.compute.manager [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 828.762470] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.855s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.762677] env[61867]: DEBUG nova.objects.instance [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lazy-loading 'resources' on Instance uuid be825f62-cb56-4647-be1c-1605536fdc89 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 829.005808] env[61867]: DEBUG oslo_concurrency.lockutils [req-715d13e6-61a3-45e3-9994-bbecf277537a req-18e41dc3-db05-4272-b6be-c914753cd3c0 service nova] Releasing lock "refresh_cache-8a83f4f2-58eb-473e-9b1e-32ce633554f9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.026968] env[61867]: DEBUG nova.network.neutron [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Successfully updated port: 104495f6-b976-4f53-9959-d5193f833ae8 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 829.045745] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Releasing lock "refresh_cache-6dda389d-a8c5-4e0e-87a5-4065e24c034e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.046082] env[61867]: DEBUG nova.compute.manager [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Instance network_info: |[{"id": "eaffa9e8-2616-4b49-9668-c8cc7ef884c0", "address": "fa:16:3e:ec:73:fc", "network": {"id": "ee84973a-7c75-4c60-8bb4-8da65e165668", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-732335052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d177558d263c4c8295ef9406bd4aee99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fa01fe1a-83b6-4c10-af75-00ddb17f9bbf", "external-id": "nsx-vlan-transportzone-431", "segmentation_id": 431, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeaffa9e8-26", "ovs_interfaceid": "eaffa9e8-2616-4b49-9668-c8cc7ef884c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 829.046832] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:73:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fa01fe1a-83b6-4c10-af75-00ddb17f9bbf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eaffa9e8-2616-4b49-9668-c8cc7ef884c0', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.059639] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Creating folder: Project (d177558d263c4c8295ef9406bd4aee99). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 829.060364] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8eae02e1-33a8-4a01-8018-f83a9729df0a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.072910] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Created folder: Project (d177558d263c4c8295ef9406bd4aee99) in parent group-v274258. [ 829.075312] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Creating folder: Instances. Parent ref: group-v274333. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 829.075940] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-961c1f88-4e1c-42ab-88d0-cd0a395660af {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.088134] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276553, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.951252} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.088134] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 6e41989e-b8fa-4009-af1e-1ce859b329a1/6e41989e-b8fa-4009-af1e-1ce859b329a1.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 829.088433] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 829.089710] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f2f79ad2-ddcc-42cf-b54e-19ceaebfebb8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.091808] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Created folder: Instances in parent group-v274333. [ 829.092111] env[61867]: DEBUG oslo.service.loopingcall [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.092668] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 829.093336] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b746dac2-99fd-49a1-9862-cce10aa6e8bb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.110292] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 829.110292] env[61867]: value = "task-1276556" [ 829.110292] env[61867]: _type = "Task" [ 829.110292] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.115052] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 829.115052] env[61867]: value = "task-1276557" [ 829.115052] env[61867]: _type = "Task" [ 829.115052] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.125301] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276556, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.130379] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276557, 'name': CreateVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.208438] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5248c0f4-5c05-0b62-14e3-2643578b4f02, 'name': SearchDatastore_Task, 'duration_secs': 0.016534} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.209323] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-034bf432-e78f-43c6-bfbb-337dbcf81809 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.215079] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 829.215079] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526cf74a-9d42-cb18-b27c-3e4f27c1dbc4" [ 829.215079] env[61867]: _type = "Task" [ 829.215079] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.223205] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526cf74a-9d42-cb18-b27c-3e4f27c1dbc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.269159] env[61867]: DEBUG nova.compute.utils [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.273914] env[61867]: DEBUG nova.compute.manager [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 829.274106] env[61867]: DEBUG nova.network.neutron [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 829.345225] env[61867]: DEBUG nova.policy [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'acee58f52c344a1abe2d0694fe55772c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2df7125ebab74e9f93520fb6b9eb3d9b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 829.497221] env[61867]: DEBUG nova.compute.manager [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Received event network-changed-eaffa9e8-2616-4b49-9668-c8cc7ef884c0 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.497221] env[61867]: DEBUG nova.compute.manager [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Refreshing instance network info cache due to event network-changed-eaffa9e8-2616-4b49-9668-c8cc7ef884c0. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 829.497221] env[61867]: DEBUG oslo_concurrency.lockutils [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] Acquiring lock "refresh_cache-6dda389d-a8c5-4e0e-87a5-4065e24c034e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.497221] env[61867]: DEBUG oslo_concurrency.lockutils [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] Acquired lock "refresh_cache-6dda389d-a8c5-4e0e-87a5-4065e24c034e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.497221] env[61867]: DEBUG nova.network.neutron [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Refreshing network info cache for port eaffa9e8-2616-4b49-9668-c8cc7ef884c0 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 829.539947] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "refresh_cache-9aab8852-addb-49e6-a59b-fa9bffc7733b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.539947] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "refresh_cache-9aab8852-addb-49e6-a59b-fa9bffc7733b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.539947] env[61867]: DEBUG nova.network.neutron [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 829.596116] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1830c4-5fb1-442d-8909-3fe1968e9d4e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.604439] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f71a74b6-4f27-493e-8bca-7486eaeec2c7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.651873] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8c2857f-f797-42c8-9922-8545f588999c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.658575] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276556, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.501238} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.663928] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 829.664409] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276557, 'name': CreateVM_Task, 'duration_secs': 0.327563} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.665332] env[61867]: DEBUG nova.network.neutron [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Successfully created port: 880e13b8-9318-4197-8bab-3bbc6ddcf952 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 829.667975] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95aa03f6-5e38-4979-b818-c05d029bbddf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.671212] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 829.672586] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d91e77b-966c-4f36-ac61-4b2ee399a90b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.677491] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.677816] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.678287] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 829.679531] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bfffb53-2e00-457e-abdc-cf4de0b43b77 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.701287] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 6e41989e-b8fa-4009-af1e-1ce859b329a1/6e41989e-b8fa-4009-af1e-1ce859b329a1.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 829.710198] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-535c3266-df5e-4d79-9339-5f061c6c4aef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.725918] env[61867]: DEBUG nova.compute.provider_tree [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.726711] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 829.726711] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52262ae1-8f1e-2b24-66c1-6aba2f8c61cc" [ 829.726711] env[61867]: _type = "Task" [ 829.726711] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.736557] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 829.736557] env[61867]: value = "task-1276558" [ 829.736557] env[61867]: _type = "Task" [ 829.736557] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.743384] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526cf74a-9d42-cb18-b27c-3e4f27c1dbc4, 'name': SearchDatastore_Task, 'duration_secs': 0.017189} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.743384] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52262ae1-8f1e-2b24-66c1-6aba2f8c61cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.746939] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.747217] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 8a83f4f2-58eb-473e-9b1e-32ce633554f9/8a83f4f2-58eb-473e-9b1e-32ce633554f9.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 829.747763] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e918f89-db0f-4303-98fc-286a7fa594fc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.757391] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.758657] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 829.758657] env[61867]: value = "task-1276559" [ 829.758657] env[61867]: _type = "Task" [ 829.758657] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.767482] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276559, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.777611] env[61867]: DEBUG nova.compute.manager [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 829.965433] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Acquiring lock "66518b2a-0242-438b-ba9f-d57c07a1165c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.965433] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Lock "66518b2a-0242-438b-ba9f-d57c07a1165c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.965433] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Acquiring lock "66518b2a-0242-438b-ba9f-d57c07a1165c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.965433] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Lock "66518b2a-0242-438b-ba9f-d57c07a1165c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.965620] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Lock "66518b2a-0242-438b-ba9f-d57c07a1165c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.966158] env[61867]: INFO nova.compute.manager [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Terminating instance [ 829.968511] env[61867]: DEBUG nova.compute.manager [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 829.968824] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 829.969738] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064a09fd-d6ab-4835-948f-b9765e291da1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.980016] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 829.980016] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e278d530-c26a-48ec-8c18-19ad24697497 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.987662] env[61867]: DEBUG oslo_vmware.api [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Waiting for the task: (returnval){ [ 829.987662] env[61867]: value = "task-1276560" [ 829.987662] env[61867]: _type = "Task" [ 829.987662] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.995536] env[61867]: DEBUG oslo_vmware.api [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': task-1276560, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.074147] env[61867]: DEBUG nova.network.neutron [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 830.228638] env[61867]: DEBUG nova.scheduler.client.report [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 830.248847] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52262ae1-8f1e-2b24-66c1-6aba2f8c61cc, 'name': SearchDatastore_Task, 'duration_secs': 0.055969} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.253691] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.254227] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.254604] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.254873] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.255290] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.256902] env[61867]: DEBUG nova.network.neutron [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Updating instance_info_cache with network_info: [{"id": "104495f6-b976-4f53-9959-d5193f833ae8", "address": "fa:16:3e:78:5e:b7", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap104495f6-b9", "ovs_interfaceid": "104495f6-b976-4f53-9959-d5193f833ae8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.259426] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b624729-76d3-4c06-bacf-41cbd1872a2c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.275233] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.288468] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276559, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.288468] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.288776] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 830.289674] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0014d0c6-50ab-4145-9264-6ed910c3d77e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.299253] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 830.299253] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5265b711-77e3-292b-1c61-efb7e46ccf42" [ 830.299253] env[61867]: _type = "Task" [ 830.299253] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.309214] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5265b711-77e3-292b-1c61-efb7e46ccf42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.463997] env[61867]: DEBUG nova.compute.manager [req-d7e3562b-44b4-43b7-86f9-fb239f5b7cec req-4e7bf608-4df3-4e43-9c4d-537b2a36b164 service nova] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Received event network-changed-f0eb4126-40e7-4fe8-b276-192b91388aba {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.464162] env[61867]: DEBUG nova.compute.manager [req-d7e3562b-44b4-43b7-86f9-fb239f5b7cec req-4e7bf608-4df3-4e43-9c4d-537b2a36b164 service nova] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Refreshing instance network info cache due to event network-changed-f0eb4126-40e7-4fe8-b276-192b91388aba. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 830.464426] env[61867]: DEBUG oslo_concurrency.lockutils [req-d7e3562b-44b4-43b7-86f9-fb239f5b7cec req-4e7bf608-4df3-4e43-9c4d-537b2a36b164 service nova] Acquiring lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.464552] env[61867]: DEBUG oslo_concurrency.lockutils [req-d7e3562b-44b4-43b7-86f9-fb239f5b7cec req-4e7bf608-4df3-4e43-9c4d-537b2a36b164 service nova] Acquired lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.464689] env[61867]: DEBUG nova.network.neutron [req-d7e3562b-44b4-43b7-86f9-fb239f5b7cec req-4e7bf608-4df3-4e43-9c4d-537b2a36b164 service nova] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Refreshing network info cache for port f0eb4126-40e7-4fe8-b276-192b91388aba {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 830.497772] env[61867]: DEBUG oslo_vmware.api [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': task-1276560, 'name': PowerOffVM_Task, 'duration_secs': 0.260899} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.498054] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 830.498209] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 830.503018] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9cf888d-57cc-4753-a398-b89e58cc1556 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.559287] env[61867]: DEBUG nova.network.neutron [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Updated VIF entry in instance network info cache for port eaffa9e8-2616-4b49-9668-c8cc7ef884c0. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 830.559607] env[61867]: DEBUG nova.network.neutron [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Updating instance_info_cache with network_info: [{"id": "eaffa9e8-2616-4b49-9668-c8cc7ef884c0", "address": "fa:16:3e:ec:73:fc", "network": {"id": "ee84973a-7c75-4c60-8bb4-8da65e165668", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-732335052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d177558d263c4c8295ef9406bd4aee99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fa01fe1a-83b6-4c10-af75-00ddb17f9bbf", "external-id": "nsx-vlan-transportzone-431", "segmentation_id": 431, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeaffa9e8-26", "ovs_interfaceid": "eaffa9e8-2616-4b49-9668-c8cc7ef884c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.569314] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 830.569393] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 830.569550] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Deleting the datastore file [datastore1] 66518b2a-0242-438b-ba9f-d57c07a1165c {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 830.569826] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-86e33b38-e57e-419f-aac4-9c566c89cdc3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.579857] env[61867]: DEBUG oslo_vmware.api [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Waiting for the task: (returnval){ [ 830.579857] env[61867]: value = "task-1276562" [ 830.579857] env[61867]: _type = "Task" [ 830.579857] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.588305] env[61867]: DEBUG oslo_vmware.api [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': task-1276562, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.740374] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.978s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.746018] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.977s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.746018] env[61867]: INFO nova.compute.claims [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 830.758691] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276558, 'name': ReconfigVM_Task, 'duration_secs': 0.72369} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.759246] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 6e41989e-b8fa-4009-af1e-1ce859b329a1/6e41989e-b8fa-4009-af1e-1ce859b329a1.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 830.759933] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f6655eb6-f2ee-40cd-86a3-e69d20b5eb3d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.766574] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "refresh_cache-9aab8852-addb-49e6-a59b-fa9bffc7733b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.766919] env[61867]: DEBUG nova.compute.manager [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Instance network_info: |[{"id": "104495f6-b976-4f53-9959-d5193f833ae8", "address": "fa:16:3e:78:5e:b7", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap104495f6-b9", "ovs_interfaceid": "104495f6-b976-4f53-9959-d5193f833ae8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 830.771393] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:5e:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f2020f39-42c4-4481-85c5-aaf03854b459', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '104495f6-b976-4f53-9959-d5193f833ae8', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.780678] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Creating folder: Project (47c5b5b783ea445eb1c02fb728b1fff2). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.781331] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 830.781331] env[61867]: value = "task-1276563" [ 830.781331] env[61867]: _type = "Task" [ 830.781331] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.782303] env[61867]: INFO nova.scheduler.client.report [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleted allocations for instance be825f62-cb56-4647-be1c-1605536fdc89 [ 830.788363] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3b3aed4b-d6d8-4071-859e-d36b5f3528e6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.793937] env[61867]: DEBUG nova.compute.manager [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 830.805214] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276559, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.682506} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.809291] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 8a83f4f2-58eb-473e-9b1e-32ce633554f9/8a83f4f2-58eb-473e-9b1e-32ce633554f9.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 830.809291] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 830.809430] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-49b0957e-fc5b-4bbc-aa95-afafef29752f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.818196] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Created folder: Project (47c5b5b783ea445eb1c02fb728b1fff2) in parent group-v274258. [ 830.820322] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Creating folder: Instances. Parent ref: group-v274336. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.820322] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276563, 'name': Rename_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.821797] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b0cc226c-dbca-4496-a55f-b469d45e37b5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.827418] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5265b711-77e3-292b-1c61-efb7e46ccf42, 'name': SearchDatastore_Task, 'duration_secs': 0.054933} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.832412] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9ab993e-d53f-4932-b1c6-7dd05c943a97 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.838182] env[61867]: DEBUG nova.virt.hardware [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 830.838182] env[61867]: DEBUG nova.virt.hardware [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 830.838490] env[61867]: DEBUG nova.virt.hardware [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 830.838582] env[61867]: DEBUG nova.virt.hardware [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 830.838794] env[61867]: DEBUG nova.virt.hardware [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 830.840428] env[61867]: DEBUG nova.virt.hardware [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 830.840428] env[61867]: DEBUG nova.virt.hardware [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 830.840428] env[61867]: DEBUG nova.virt.hardware [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 830.840428] env[61867]: DEBUG nova.virt.hardware [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 830.840428] env[61867]: DEBUG nova.virt.hardware [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 830.840941] env[61867]: DEBUG nova.virt.hardware [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 830.841960] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b943115a-e6b6-4042-9271-d716a15cc994 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.849179] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 830.849179] env[61867]: value = "task-1276566" [ 830.849179] env[61867]: _type = "Task" [ 830.849179] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.849179] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Created folder: Instances in parent group-v274336. [ 830.849179] env[61867]: DEBUG oslo.service.loopingcall [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.849725] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 830.853024] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 830.853024] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5203dfa9-ffa2-3c2d-9780-39a7d4505871" [ 830.853024] env[61867]: _type = "Task" [ 830.853024] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.853319] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dcac0165-5a04-423c-8c6f-ba420e577cff {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.877725] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4986e165-33ce-4899-80fd-9890e5a53384 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.886173] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.886173] env[61867]: value = "task-1276567" [ 830.886173] env[61867]: _type = "Task" [ 830.886173] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.888873] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276566, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.901498] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5203dfa9-ffa2-3c2d-9780-39a7d4505871, 'name': SearchDatastore_Task, 'duration_secs': 0.011012} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.902550] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.903567] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 6dda389d-a8c5-4e0e-87a5-4065e24c034e/6dda389d-a8c5-4e0e-87a5-4065e24c034e.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 830.903567] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3942f9d1-0cec-4dfd-80bb-a9af34f1af50 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.907873] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276567, 'name': CreateVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.912521] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 830.912521] env[61867]: value = "task-1276568" [ 830.912521] env[61867]: _type = "Task" [ 830.912521] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.921995] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276568, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.936787] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.937030] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.062350] env[61867]: DEBUG oslo_concurrency.lockutils [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] Releasing lock "refresh_cache-6dda389d-a8c5-4e0e-87a5-4065e24c034e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.062661] env[61867]: DEBUG nova.compute.manager [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Received event network-vif-plugged-104495f6-b976-4f53-9959-d5193f833ae8 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.062855] env[61867]: DEBUG oslo_concurrency.lockutils [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] Acquiring lock "9aab8852-addb-49e6-a59b-fa9bffc7733b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.063167] env[61867]: DEBUG oslo_concurrency.lockutils [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] Lock "9aab8852-addb-49e6-a59b-fa9bffc7733b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.063412] env[61867]: DEBUG oslo_concurrency.lockutils [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] Lock "9aab8852-addb-49e6-a59b-fa9bffc7733b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.063662] env[61867]: DEBUG nova.compute.manager [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] No waiting events found dispatching network-vif-plugged-104495f6-b976-4f53-9959-d5193f833ae8 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 831.063966] env[61867]: WARNING nova.compute.manager [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Received unexpected event network-vif-plugged-104495f6-b976-4f53-9959-d5193f833ae8 for instance with vm_state building and task_state spawning. [ 831.063966] env[61867]: DEBUG nova.compute.manager [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Received event network-changed-104495f6-b976-4f53-9959-d5193f833ae8 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.064151] env[61867]: DEBUG nova.compute.manager [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Refreshing instance network info cache due to event network-changed-104495f6-b976-4f53-9959-d5193f833ae8. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 831.064353] env[61867]: DEBUG oslo_concurrency.lockutils [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] Acquiring lock "refresh_cache-9aab8852-addb-49e6-a59b-fa9bffc7733b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.064492] env[61867]: DEBUG oslo_concurrency.lockutils [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] Acquired lock "refresh_cache-9aab8852-addb-49e6-a59b-fa9bffc7733b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.064701] env[61867]: DEBUG nova.network.neutron [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Refreshing network info cache for port 104495f6-b976-4f53-9959-d5193f833ae8 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 831.089419] env[61867]: DEBUG oslo_vmware.api [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Task: {'id': task-1276562, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162912} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.089587] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 831.089773] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 831.089954] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 831.090525] env[61867]: INFO nova.compute.manager [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 831.090731] env[61867]: DEBUG oslo.service.loopingcall [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.091308] env[61867]: DEBUG nova.compute.manager [-] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 831.091308] env[61867]: DEBUG nova.network.neutron [-] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 831.294114] env[61867]: DEBUG oslo_concurrency.lockutils [None req-7fb80d6c-17ea-4d2d-aad5-184c86678721 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "be825f62-cb56-4647-be1c-1605536fdc89" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.437s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.312387] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276563, 'name': Rename_Task, 'duration_secs': 0.171097} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.312905] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 831.313315] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9438aceb-9989-4e8d-8058-fd4fe77a6f19 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.322418] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 831.322418] env[61867]: value = "task-1276569" [ 831.322418] env[61867]: _type = "Task" [ 831.322418] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.332740] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276569, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.359853] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276566, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079595} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.360159] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 831.361368] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a05917-75d4-4d4f-bc96-84df66006b88 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.390019] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 8a83f4f2-58eb-473e-9b1e-32ce633554f9/8a83f4f2-58eb-473e-9b1e-32ce633554f9.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 831.390709] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27fa28e9-10f7-485d-b48e-15be348d3484 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.407310] env[61867]: DEBUG nova.network.neutron [req-d7e3562b-44b4-43b7-86f9-fb239f5b7cec req-4e7bf608-4df3-4e43-9c4d-537b2a36b164 service nova] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updated VIF entry in instance network info cache for port f0eb4126-40e7-4fe8-b276-192b91388aba. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 831.407779] env[61867]: DEBUG nova.network.neutron [req-d7e3562b-44b4-43b7-86f9-fb239f5b7cec req-4e7bf608-4df3-4e43-9c4d-537b2a36b164 service nova] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updating instance_info_cache with network_info: [{"id": "f0eb4126-40e7-4fe8-b276-192b91388aba", "address": "fa:16:3e:69:84:03", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0eb4126-40", "ovs_interfaceid": "f0eb4126-40e7-4fe8-b276-192b91388aba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.420147] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276567, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.422784] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 831.422784] env[61867]: value = "task-1276570" [ 831.422784] env[61867]: _type = "Task" [ 831.422784] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.429888] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276568, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.439184] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.557834] env[61867]: DEBUG nova.compute.manager [req-4918fcf8-1b3a-4ba3-ba83-0e34f7ed1db7 req-2ef429fe-253e-4da3-847b-aa691e1daae6 service nova] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Received event network-vif-plugged-880e13b8-9318-4197-8bab-3bbc6ddcf952 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.558090] env[61867]: DEBUG oslo_concurrency.lockutils [req-4918fcf8-1b3a-4ba3-ba83-0e34f7ed1db7 req-2ef429fe-253e-4da3-847b-aa691e1daae6 service nova] Acquiring lock "ad699b24-d01d-4d7a-815f-c6b10286012d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.558591] env[61867]: DEBUG oslo_concurrency.lockutils [req-4918fcf8-1b3a-4ba3-ba83-0e34f7ed1db7 req-2ef429fe-253e-4da3-847b-aa691e1daae6 service nova] Lock "ad699b24-d01d-4d7a-815f-c6b10286012d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.558629] env[61867]: DEBUG oslo_concurrency.lockutils [req-4918fcf8-1b3a-4ba3-ba83-0e34f7ed1db7 req-2ef429fe-253e-4da3-847b-aa691e1daae6 service nova] Lock "ad699b24-d01d-4d7a-815f-c6b10286012d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.558793] env[61867]: DEBUG nova.compute.manager [req-4918fcf8-1b3a-4ba3-ba83-0e34f7ed1db7 req-2ef429fe-253e-4da3-847b-aa691e1daae6 service nova] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] No waiting events found dispatching network-vif-plugged-880e13b8-9318-4197-8bab-3bbc6ddcf952 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 831.559091] env[61867]: WARNING nova.compute.manager [req-4918fcf8-1b3a-4ba3-ba83-0e34f7ed1db7 req-2ef429fe-253e-4da3-847b-aa691e1daae6 service nova] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Received unexpected event network-vif-plugged-880e13b8-9318-4197-8bab-3bbc6ddcf952 for instance with vm_state building and task_state spawning. [ 831.708804] env[61867]: DEBUG nova.network.neutron [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Successfully updated port: 880e13b8-9318-4197-8bab-3bbc6ddcf952 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 831.844841] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276569, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.900949] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276567, 'name': CreateVM_Task, 'duration_secs': 0.593527} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.902207] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 831.902207] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.902207] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.902733] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 831.903258] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfef84e0-01b1-41ef-a594-c74f81d12085 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.914452] env[61867]: DEBUG oslo_concurrency.lockutils [req-d7e3562b-44b4-43b7-86f9-fb239f5b7cec req-4e7bf608-4df3-4e43-9c4d-537b2a36b164 service nova] Releasing lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.915008] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 831.915008] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529a7d37-4078-8851-1c62-40e17745904c" [ 831.915008] env[61867]: _type = "Task" [ 831.915008] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.941050] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276568, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.634} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.945168] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 6dda389d-a8c5-4e0e-87a5-4065e24c034e/6dda389d-a8c5-4e0e-87a5-4065e24c034e.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 831.945426] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 831.945704] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529a7d37-4078-8851-1c62-40e17745904c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.945918] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ae3d47b6-74b2-436f-ae9e-3fb24fdfa560 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.950592] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276570, 'name': ReconfigVM_Task, 'duration_secs': 0.33896} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.951200] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 8a83f4f2-58eb-473e-9b1e-32ce633554f9/8a83f4f2-58eb-473e-9b1e-32ce633554f9.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 831.951832] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ccb605e6-7b68-4baf-a585-2e678b240c25 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.955147] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 831.955147] env[61867]: value = "task-1276571" [ 831.955147] env[61867]: _type = "Task" [ 831.955147] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.961117] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 831.961117] env[61867]: value = "task-1276572" [ 831.961117] env[61867]: _type = "Task" [ 831.961117] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.967696] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276571, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.972474] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276572, 'name': Rename_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.031652] env[61867]: DEBUG nova.network.neutron [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Updated VIF entry in instance network info cache for port 104495f6-b976-4f53-9959-d5193f833ae8. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 832.032034] env[61867]: DEBUG nova.network.neutron [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Updating instance_info_cache with network_info: [{"id": "104495f6-b976-4f53-9959-d5193f833ae8", "address": "fa:16:3e:78:5e:b7", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap104495f6-b9", "ovs_interfaceid": "104495f6-b976-4f53-9959-d5193f833ae8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.132105] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3470b8d-547c-4728-8c97-c382bfa7f48d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.141458] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ce0757-ce87-4fb9-9025-a31add8bbe1a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.177683] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80809879-60c7-4fb2-9a7a-65ee8ec1904e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.180383] env[61867]: DEBUG nova.network.neutron [-] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.186750] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b0b087-1076-45b5-a8a9-7bc1c48488bb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.200657] env[61867]: DEBUG nova.compute.provider_tree [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.210144] env[61867]: DEBUG oslo_concurrency.lockutils [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Acquiring lock "refresh_cache-ad699b24-d01d-4d7a-815f-c6b10286012d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.210302] env[61867]: DEBUG oslo_concurrency.lockutils [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Acquired lock "refresh_cache-ad699b24-d01d-4d7a-815f-c6b10286012d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.210487] env[61867]: DEBUG nova.network.neutron [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 832.336470] env[61867]: DEBUG oslo_vmware.api [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276569, 'name': PowerOnVM_Task, 'duration_secs': 0.785016} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.336752] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 832.337216] env[61867]: INFO nova.compute.manager [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Took 11.24 seconds to spawn the instance on the hypervisor. [ 832.337411] env[61867]: DEBUG nova.compute.manager [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 832.338206] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6c82e8-cc23-4ebe-bc64-28bfc5175c3c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.430232] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529a7d37-4078-8851-1c62-40e17745904c, 'name': SearchDatastore_Task, 'duration_secs': 0.054401} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.430232] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.430232] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 832.430232] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.430521] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.430521] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 832.430521] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aa2f2c16-4a4f-4852-88fe-a584e1099183 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.445867] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.445931] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 832.446684] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0acdac5-3f65-489e-b299-a586253f9f73 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.455201] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 832.455201] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b615df-a6a3-9a72-e11a-2b29d9a34b5c" [ 832.455201] env[61867]: _type = "Task" [ 832.455201] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.472309] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b615df-a6a3-9a72-e11a-2b29d9a34b5c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.475339] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276571, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107221} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.475591] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 832.476347] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc810bae-89d0-45ae-affa-903fa6e90af6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.481931] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276572, 'name': Rename_Task, 'duration_secs': 0.154414} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.482486] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 832.482775] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f04ddb36-f657-472a-9d12-4dbb31a0e2e2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.501518] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 6dda389d-a8c5-4e0e-87a5-4065e24c034e/6dda389d-a8c5-4e0e-87a5-4065e24c034e.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 832.502265] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-647bdd84-ade2-4642-bd6e-f4710792f61d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.518855] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 832.518855] env[61867]: value = "task-1276573" [ 832.518855] env[61867]: _type = "Task" [ 832.518855] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.524272] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 832.524272] env[61867]: value = "task-1276574" [ 832.524272] env[61867]: _type = "Task" [ 832.524272] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.531480] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276573, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.536903] env[61867]: DEBUG oslo_concurrency.lockutils [req-d7f9377b-dc4d-4b03-90c2-58dd7276948a req-cd409c23-4ab2-4c69-99d2-445352ab2a55 service nova] Releasing lock "refresh_cache-9aab8852-addb-49e6-a59b-fa9bffc7733b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.537687] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276574, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.682628] env[61867]: INFO nova.compute.manager [-] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Took 1.59 seconds to deallocate network for instance. [ 832.703834] env[61867]: DEBUG nova.scheduler.client.report [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 832.765765] env[61867]: DEBUG nova.network.neutron [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.856720] env[61867]: INFO nova.compute.manager [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Took 47.02 seconds to build instance. [ 832.967071] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b615df-a6a3-9a72-e11a-2b29d9a34b5c, 'name': SearchDatastore_Task, 'duration_secs': 0.071215} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.968339] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e538009-de25-4eed-835a-f7b733a6ec7c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.971998] env[61867]: DEBUG nova.network.neutron [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Updating instance_info_cache with network_info: [{"id": "880e13b8-9318-4197-8bab-3bbc6ddcf952", "address": "fa:16:3e:bf:35:64", "network": {"id": "35638bd0-a0d5-4279-8967-9e5bf5d18b03", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1566301482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2df7125ebab74e9f93520fb6b9eb3d9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap880e13b8-93", "ovs_interfaceid": "880e13b8-9318-4197-8bab-3bbc6ddcf952", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.976284] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 832.976284] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]524b9b33-75b6-d132-bf2b-dd7a765b44bb" [ 832.976284] env[61867]: _type = "Task" [ 832.976284] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.984062] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]524b9b33-75b6-d132-bf2b-dd7a765b44bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.003865] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "9c23a44c-eb72-4194-a3e5-88a8ef54ed24" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.004163] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "9c23a44c-eb72-4194-a3e5-88a8ef54ed24" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.033274] env[61867]: DEBUG oslo_vmware.api [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276573, 'name': PowerOnVM_Task, 'duration_secs': 0.483209} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.036821] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 833.036821] env[61867]: INFO nova.compute.manager [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Took 9.48 seconds to spawn the instance on the hypervisor. [ 833.036983] env[61867]: DEBUG nova.compute.manager [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.037588] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276574, 'name': ReconfigVM_Task, 'duration_secs': 0.347338} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.038282] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3abe0a1f-361c-47ad-bfc4-f3847f2ad48c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.040809] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 6dda389d-a8c5-4e0e-87a5-4065e24c034e/6dda389d-a8c5-4e0e-87a5-4065e24c034e.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 833.041574] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c602e197-6daf-4b6d-92a8-abe8b3688249 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.052030] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 833.052030] env[61867]: value = "task-1276575" [ 833.052030] env[61867]: _type = "Task" [ 833.052030] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.060617] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276575, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.192552] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.210633] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.467s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.211214] env[61867]: DEBUG nova.compute.manager [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 833.214130] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.285s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.215862] env[61867]: INFO nova.compute.claims [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 833.359772] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd6ffbd3-ba34-4d02-8247-857652f31a7f tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "6e41989e-b8fa-4009-af1e-1ce859b329a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.020s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.474958] env[61867]: DEBUG oslo_concurrency.lockutils [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Releasing lock "refresh_cache-ad699b24-d01d-4d7a-815f-c6b10286012d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.475320] env[61867]: DEBUG nova.compute.manager [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Instance network_info: |[{"id": "880e13b8-9318-4197-8bab-3bbc6ddcf952", "address": "fa:16:3e:bf:35:64", "network": {"id": "35638bd0-a0d5-4279-8967-9e5bf5d18b03", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1566301482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2df7125ebab74e9f93520fb6b9eb3d9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap880e13b8-93", "ovs_interfaceid": "880e13b8-9318-4197-8bab-3bbc6ddcf952", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 833.475759] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:35:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10ff2092-e8eb-4768-ad4a-65a80560b447', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '880e13b8-9318-4197-8bab-3bbc6ddcf952', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.483358] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Creating folder: Project (2df7125ebab74e9f93520fb6b9eb3d9b). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.483738] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-faeeed0f-646e-4f15-bca8-a6fdae6fedfe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.494601] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]524b9b33-75b6-d132-bf2b-dd7a765b44bb, 'name': SearchDatastore_Task, 'duration_secs': 0.0145} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.494844] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.495106] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 9aab8852-addb-49e6-a59b-fa9bffc7733b/9aab8852-addb-49e6-a59b-fa9bffc7733b.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 833.495354] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-573258e0-156e-40f0-abaa-f444d44ede40 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.498018] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Created folder: Project (2df7125ebab74e9f93520fb6b9eb3d9b) in parent group-v274258. [ 833.498217] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Creating folder: Instances. Parent ref: group-v274339. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.498723] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a11314b7-83dc-4869-8bdf-1d22bbc3d345 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.502539] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 833.502539] env[61867]: value = "task-1276577" [ 833.502539] env[61867]: _type = "Task" [ 833.502539] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.508205] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Created folder: Instances in parent group-v274339. [ 833.508428] env[61867]: DEBUG oslo.service.loopingcall [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.511137] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 833.511391] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276577, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.511588] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-93ca305d-d4c3-48e6-8919-dd03fab8b340 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.529348] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.529348] env[61867]: value = "task-1276579" [ 833.529348] env[61867]: _type = "Task" [ 833.529348] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.536342] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276579, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.563573] env[61867]: INFO nova.compute.manager [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Took 38.65 seconds to build instance. [ 833.568029] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276575, 'name': Rename_Task, 'duration_secs': 0.237728} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.568653] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 833.568779] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-05cad9bb-6481-4a3b-bdc2-72eb177cbe0b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.575357] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 833.575357] env[61867]: value = "task-1276580" [ 833.575357] env[61867]: _type = "Task" [ 833.575357] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.584037] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276580, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.721198] env[61867]: DEBUG nova.compute.utils [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 833.725722] env[61867]: DEBUG nova.compute.manager [req-984ebe6a-1e31-4708-9d2e-332b7204e41d req-07aa3311-c57c-4a80-87cb-c66878db6b27 service nova] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Received event network-vif-deleted-82c63e6a-f20c-460d-938c-d7af7e99764c {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.725942] env[61867]: DEBUG nova.compute.manager [req-984ebe6a-1e31-4708-9d2e-332b7204e41d req-07aa3311-c57c-4a80-87cb-c66878db6b27 service nova] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Received event network-changed-880e13b8-9318-4197-8bab-3bbc6ddcf952 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.727214] env[61867]: DEBUG nova.compute.manager [req-984ebe6a-1e31-4708-9d2e-332b7204e41d req-07aa3311-c57c-4a80-87cb-c66878db6b27 service nova] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Refreshing instance network info cache due to event network-changed-880e13b8-9318-4197-8bab-3bbc6ddcf952. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 833.727519] env[61867]: DEBUG oslo_concurrency.lockutils [req-984ebe6a-1e31-4708-9d2e-332b7204e41d req-07aa3311-c57c-4a80-87cb-c66878db6b27 service nova] Acquiring lock "refresh_cache-ad699b24-d01d-4d7a-815f-c6b10286012d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.727700] env[61867]: DEBUG oslo_concurrency.lockutils [req-984ebe6a-1e31-4708-9d2e-332b7204e41d req-07aa3311-c57c-4a80-87cb-c66878db6b27 service nova] Acquired lock "refresh_cache-ad699b24-d01d-4d7a-815f-c6b10286012d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.728720] env[61867]: DEBUG nova.network.neutron [req-984ebe6a-1e31-4708-9d2e-332b7204e41d req-07aa3311-c57c-4a80-87cb-c66878db6b27 service nova] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Refreshing network info cache for port 880e13b8-9318-4197-8bab-3bbc6ddcf952 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 833.729706] env[61867]: DEBUG nova.compute.manager [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 833.729880] env[61867]: DEBUG nova.network.neutron [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 833.800096] env[61867]: DEBUG nova.policy [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ea634246436422a9407f82692b69ede', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '86ced20eaf4740e298dc6f8ca5550c09', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 833.862302] env[61867]: DEBUG nova.compute.manager [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 834.018394] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276577, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.042332] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276579, 'name': CreateVM_Task, 'duration_secs': 0.34696} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.042527] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 834.043303] env[61867]: DEBUG oslo_concurrency.lockutils [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.043584] env[61867]: DEBUG oslo_concurrency.lockutils [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.043831] env[61867]: DEBUG oslo_concurrency.lockutils [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.044132] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26556766-7c6f-4c06-889d-7bfe4b3a3bd6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.050393] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Waiting for the task: (returnval){ [ 834.050393] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d15174-4160-349d-5632-e89a4216e1ec" [ 834.050393] env[61867]: _type = "Task" [ 834.050393] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.059119] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d15174-4160-349d-5632-e89a4216e1ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.069854] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9b1a9119-510a-4b8b-bbf2-6647a31ad9cf tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "8a83f4f2-58eb-473e-9b1e-32ce633554f9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.238s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.085806] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276580, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.230752] env[61867]: DEBUG nova.compute.manager [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 834.382013] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.410923] env[61867]: DEBUG nova.network.neutron [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Successfully created port: af89c243-c119-47c9-9eda-60c9be2b8c51 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 834.451761] env[61867]: INFO nova.compute.manager [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Rescuing [ 834.453023] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "refresh_cache-8a83f4f2-58eb-473e-9b1e-32ce633554f9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.453023] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired lock "refresh_cache-8a83f4f2-58eb-473e-9b1e-32ce633554f9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.453023] env[61867]: DEBUG nova.network.neutron [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 834.454273] env[61867]: DEBUG nova.network.neutron [req-984ebe6a-1e31-4708-9d2e-332b7204e41d req-07aa3311-c57c-4a80-87cb-c66878db6b27 service nova] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Updated VIF entry in instance network info cache for port 880e13b8-9318-4197-8bab-3bbc6ddcf952. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 834.455182] env[61867]: DEBUG nova.network.neutron [req-984ebe6a-1e31-4708-9d2e-332b7204e41d req-07aa3311-c57c-4a80-87cb-c66878db6b27 service nova] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Updating instance_info_cache with network_info: [{"id": "880e13b8-9318-4197-8bab-3bbc6ddcf952", "address": "fa:16:3e:bf:35:64", "network": {"id": "35638bd0-a0d5-4279-8967-9e5bf5d18b03", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1566301482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2df7125ebab74e9f93520fb6b9eb3d9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap880e13b8-93", "ovs_interfaceid": "880e13b8-9318-4197-8bab-3bbc6ddcf952", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.513525] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276577, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.550242} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.516188] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 9aab8852-addb-49e6-a59b-fa9bffc7733b/9aab8852-addb-49e6-a59b-fa9bffc7733b.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 834.516444] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.517155] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6143840b-bb09-4e36-b96c-4afa77adfe77 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.523828] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 834.523828] env[61867]: value = "task-1276581" [ 834.523828] env[61867]: _type = "Task" [ 834.523828] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.532697] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276581, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.547185] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a0da79a-2db8-4ebf-8a5e-625a10bf7f3d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.559958] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0de4a54a-9b07-4072-a856-3bb193fa14a0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.569112] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d15174-4160-349d-5632-e89a4216e1ec, 'name': SearchDatastore_Task, 'duration_secs': 0.012342} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.597672] env[61867]: DEBUG oslo_concurrency.lockutils [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.598426] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.598426] env[61867]: DEBUG oslo_concurrency.lockutils [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.598426] env[61867]: DEBUG oslo_concurrency.lockutils [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.598581] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.598896] env[61867]: DEBUG nova.compute.manager [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 834.602824] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3e720fd4-8204-4b94-83e4-f4cfb0e44269 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.608628] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9803c11b-9dc9-40fc-b3fe-9ffcb5767853 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.620495] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abf37b8f-5867-4182-96aa-2cf6c4aaa968 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.628176] env[61867]: DEBUG oslo_vmware.api [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276580, 'name': PowerOnVM_Task, 'duration_secs': 0.614873} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.628176] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 834.628176] env[61867]: INFO nova.compute.manager [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Took 8.64 seconds to spawn the instance on the hypervisor. [ 834.628176] env[61867]: DEBUG nova.compute.manager [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 834.628176] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.629174] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 834.629174] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9059cbbf-eacc-4f32-9b05-eef6391b053a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.631436] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb96069b-fe79-44aa-b4d9-86f42ff023c2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.643959] env[61867]: DEBUG nova.compute.provider_tree [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.652277] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Waiting for the task: (returnval){ [ 834.652277] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521abe0b-7b9d-40d8-fad5-f9a62cc950c6" [ 834.652277] env[61867]: _type = "Task" [ 834.652277] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.661141] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521abe0b-7b9d-40d8-fad5-f9a62cc950c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.959579] env[61867]: DEBUG oslo_concurrency.lockutils [req-984ebe6a-1e31-4708-9d2e-332b7204e41d req-07aa3311-c57c-4a80-87cb-c66878db6b27 service nova] Releasing lock "refresh_cache-ad699b24-d01d-4d7a-815f-c6b10286012d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.035817] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276581, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.326602} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.036019] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 835.036829] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c40f922a-0655-4057-9b01-bad4641f27a7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.060080] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 9aab8852-addb-49e6-a59b-fa9bffc7733b/9aab8852-addb-49e6-a59b-fa9bffc7733b.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.060367] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b10a0d4-520e-4580-9dda-c6559e158d92 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.086021] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 835.086021] env[61867]: value = "task-1276582" [ 835.086021] env[61867]: _type = "Task" [ 835.086021] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.096159] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276582, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.119564] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.146336] env[61867]: DEBUG nova.scheduler.client.report [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 835.166474] env[61867]: INFO nova.compute.manager [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Took 38.32 seconds to build instance. [ 835.175341] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521abe0b-7b9d-40d8-fad5-f9a62cc950c6, 'name': SearchDatastore_Task, 'duration_secs': 0.021224} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.176414] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73df5087-4c35-4be5-a609-6f179833d0fe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.182617] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Waiting for the task: (returnval){ [ 835.182617] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52260b64-f661-2f81-0f4c-312d3a6596cc" [ 835.182617] env[61867]: _type = "Task" [ 835.182617] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.192121] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52260b64-f661-2f81-0f4c-312d3a6596cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.246304] env[61867]: DEBUG nova.compute.manager [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 835.249244] env[61867]: DEBUG nova.network.neutron [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Updating instance_info_cache with network_info: [{"id": "cdf39635-188f-4031-ae51-c7d98a8e8365", "address": "fa:16:3e:01:31:7c", "network": {"id": "ab6ec5b9-8193-43c9-a3ce-d63f31e5a001", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1573384141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35aa85105b394369b17ef4ce51a49066", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdf39635-18", "ovs_interfaceid": "cdf39635-188f-4031-ae51-c7d98a8e8365", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.276732] env[61867]: DEBUG nova.virt.hardware [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 835.277012] env[61867]: DEBUG nova.virt.hardware [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 835.277573] env[61867]: DEBUG nova.virt.hardware [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.277810] env[61867]: DEBUG nova.virt.hardware [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 835.277964] env[61867]: DEBUG nova.virt.hardware [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.278127] env[61867]: DEBUG nova.virt.hardware [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 835.278336] env[61867]: DEBUG nova.virt.hardware [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 835.278490] env[61867]: DEBUG nova.virt.hardware [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 835.278653] env[61867]: DEBUG nova.virt.hardware [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 835.278814] env[61867]: DEBUG nova.virt.hardware [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 835.278985] env[61867]: DEBUG nova.virt.hardware [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.280081] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9b668a6-bc80-4ae8-a737-d98c7995a7c2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.288111] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee6f4388-9485-4539-8b4b-fa8d94f6a5ab {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.593891] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276582, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.651479] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.652107] env[61867]: DEBUG nova.compute.manager [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 835.655361] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 20.339s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.669847] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92671df5-69e2-478b-a794-56f5c17ad17d tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.851s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.695052] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52260b64-f661-2f81-0f4c-312d3a6596cc, 'name': SearchDatastore_Task, 'duration_secs': 0.017484} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.695168] env[61867]: DEBUG oslo_concurrency.lockutils [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.695386] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] ad699b24-d01d-4d7a-815f-c6b10286012d/ad699b24-d01d-4d7a-815f-c6b10286012d.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 835.695661] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-23f024b1-6d7d-482f-83b0-08e63619d8b2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.704129] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Waiting for the task: (returnval){ [ 835.704129] env[61867]: value = "task-1276583" [ 835.704129] env[61867]: _type = "Task" [ 835.704129] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.712965] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': task-1276583, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.752864] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Releasing lock "refresh_cache-8a83f4f2-58eb-473e-9b1e-32ce633554f9" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.097110] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276582, 'name': ReconfigVM_Task, 'duration_secs': 0.779659} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.098705] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 9aab8852-addb-49e6-a59b-fa9bffc7733b/9aab8852-addb-49e6-a59b-fa9bffc7733b.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.100509] env[61867]: DEBUG nova.compute.manager [req-9798b587-02c2-496a-9630-4cba62ec3c3c req-a182f260-ffec-45a8-8bd9-9ad6ef08f9a0 service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Received event network-vif-plugged-af89c243-c119-47c9-9eda-60c9be2b8c51 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.100782] env[61867]: DEBUG oslo_concurrency.lockutils [req-9798b587-02c2-496a-9630-4cba62ec3c3c req-a182f260-ffec-45a8-8bd9-9ad6ef08f9a0 service nova] Acquiring lock "ca2f6d18-f773-4875-83de-2f2be912c2f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.101029] env[61867]: DEBUG oslo_concurrency.lockutils [req-9798b587-02c2-496a-9630-4cba62ec3c3c req-a182f260-ffec-45a8-8bd9-9ad6ef08f9a0 service nova] Lock "ca2f6d18-f773-4875-83de-2f2be912c2f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.101212] env[61867]: DEBUG oslo_concurrency.lockutils [req-9798b587-02c2-496a-9630-4cba62ec3c3c req-a182f260-ffec-45a8-8bd9-9ad6ef08f9a0 service nova] Lock "ca2f6d18-f773-4875-83de-2f2be912c2f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.101388] env[61867]: DEBUG nova.compute.manager [req-9798b587-02c2-496a-9630-4cba62ec3c3c req-a182f260-ffec-45a8-8bd9-9ad6ef08f9a0 service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] No waiting events found dispatching network-vif-plugged-af89c243-c119-47c9-9eda-60c9be2b8c51 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 836.101613] env[61867]: WARNING nova.compute.manager [req-9798b587-02c2-496a-9630-4cba62ec3c3c req-a182f260-ffec-45a8-8bd9-9ad6ef08f9a0 service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Received unexpected event network-vif-plugged-af89c243-c119-47c9-9eda-60c9be2b8c51 for instance with vm_state building and task_state spawning. [ 836.102010] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e9fe567f-5982-4a18-8a0f-1420ea46e463 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.110370] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 836.110370] env[61867]: value = "task-1276584" [ 836.110370] env[61867]: _type = "Task" [ 836.110370] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.121778] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276584, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.158733] env[61867]: DEBUG nova.compute.utils [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 836.161292] env[61867]: DEBUG nova.compute.manager [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 836.161575] env[61867]: DEBUG nova.network.neutron [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 836.168436] env[61867]: INFO nova.compute.claims [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.214726] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': task-1276583, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.284615] env[61867]: DEBUG nova.policy [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '808c75aeb22e407c831849555144ab63', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e54d4ce9d6c24e8ab40cb77886aac634', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 836.289922] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 836.290284] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31ce6b17-da49-408d-98d9-926af254d36f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.300461] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 836.300461] env[61867]: value = "task-1276585" [ 836.300461] env[61867]: _type = "Task" [ 836.300461] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.317196] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276585, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.321546] env[61867]: DEBUG nova.network.neutron [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Successfully updated port: af89c243-c119-47c9-9eda-60c9be2b8c51 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 836.620323] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276584, 'name': Rename_Task, 'duration_secs': 0.348751} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.620699] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 836.621012] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c76aa48b-4ab3-4cd2-a491-be486e21169e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.628412] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 836.628412] env[61867]: value = "task-1276586" [ 836.628412] env[61867]: _type = "Task" [ 836.628412] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.641645] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276586, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.674752] env[61867]: DEBUG nova.compute.manager [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 836.680690] env[61867]: INFO nova.compute.resource_tracker [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Updating resource usage from migration 304235dc-4d01-4bfd-af03-d58b4a8ac04a [ 836.713626] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': task-1276583, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.635879} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.713946] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] ad699b24-d01d-4d7a-815f-c6b10286012d/ad699b24-d01d-4d7a-815f-c6b10286012d.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 836.714559] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 836.714559] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c2daaa18-ae8c-490a-9727-7c59964025a5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.720713] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Waiting for the task: (returnval){ [ 836.720713] env[61867]: value = "task-1276587" [ 836.720713] env[61867]: _type = "Task" [ 836.720713] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.734525] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': task-1276587, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.815846] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276585, 'name': PowerOffVM_Task, 'duration_secs': 0.318754} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.817451] env[61867]: DEBUG nova.network.neutron [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Successfully created port: cf5eac56-8386-49e1-a2b3-d19e3d8297d6 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 836.822099] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 836.823174] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb0e25f3-f43f-4f02-bf0f-7fcf9143ab0b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.828144] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.828144] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.828144] env[61867]: DEBUG nova.network.neutron [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 836.855989] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc237ae-bb72-462c-a233-ed2295abfca7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.870934] env[61867]: DEBUG nova.compute.manager [req-bfcdb9a4-ffc4-4d08-b259-87d3cc695bf6 req-868e0028-8250-4e53-8ec5-43c42fbbd30f service nova] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Received event network-changed-eaffa9e8-2616-4b49-9668-c8cc7ef884c0 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.871210] env[61867]: DEBUG nova.compute.manager [req-bfcdb9a4-ffc4-4d08-b259-87d3cc695bf6 req-868e0028-8250-4e53-8ec5-43c42fbbd30f service nova] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Refreshing instance network info cache due to event network-changed-eaffa9e8-2616-4b49-9668-c8cc7ef884c0. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 836.871358] env[61867]: DEBUG oslo_concurrency.lockutils [req-bfcdb9a4-ffc4-4d08-b259-87d3cc695bf6 req-868e0028-8250-4e53-8ec5-43c42fbbd30f service nova] Acquiring lock "refresh_cache-6dda389d-a8c5-4e0e-87a5-4065e24c034e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.871443] env[61867]: DEBUG oslo_concurrency.lockutils [req-bfcdb9a4-ffc4-4d08-b259-87d3cc695bf6 req-868e0028-8250-4e53-8ec5-43c42fbbd30f service nova] Acquired lock "refresh_cache-6dda389d-a8c5-4e0e-87a5-4065e24c034e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.871633] env[61867]: DEBUG nova.network.neutron [req-bfcdb9a4-ffc4-4d08-b259-87d3cc695bf6 req-868e0028-8250-4e53-8ec5-43c42fbbd30f service nova] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Refreshing network info cache for port eaffa9e8-2616-4b49-9668-c8cc7ef884c0 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 836.894103] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 836.894784] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0802e631-77c5-4f57-a0c0-71eefc417cc4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.901856] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 836.901856] env[61867]: value = "task-1276588" [ 836.901856] env[61867]: _type = "Task" [ 836.901856] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.913012] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] VM already powered off {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 836.913262] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.913503] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.913674] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.913875] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.914132] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6282dd75-0fc0-4234-b5bb-873f605abcf1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.922498] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.922658] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 836.923409] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa258a67-a480-4621-8667-2d8f481f411e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.931323] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 836.931323] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528838c0-8e26-af05-10f2-ae17acbebe52" [ 836.931323] env[61867]: _type = "Task" [ 836.931323] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.939328] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528838c0-8e26-af05-10f2-ae17acbebe52, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.987136] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0776b322-479a-408a-9bd7-eea9af76e610 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.996680] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae0dcb1-9bae-42de-ac06-ec4806c4d391 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.034187] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202a693d-f04b-4aa6-a840-dc426c94c8ed {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.040197] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7789477d-f126-4259-a30a-3c09aee4be79 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.054902] env[61867]: DEBUG nova.compute.provider_tree [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.138923] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276586, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.231539] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': task-1276587, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064448} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.231856] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 837.232717] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba761a90-206d-4c12-8b9a-3275ea9573ac {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.256734] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] ad699b24-d01d-4d7a-815f-c6b10286012d/ad699b24-d01d-4d7a-815f-c6b10286012d.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.257145] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07c72581-7e4d-4aa1-892d-3904cfcbbbbd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.277564] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Waiting for the task: (returnval){ [ 837.277564] env[61867]: value = "task-1276589" [ 837.277564] env[61867]: _type = "Task" [ 837.277564] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.286893] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': task-1276589, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.436229] env[61867]: DEBUG nova.network.neutron [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 837.444258] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528838c0-8e26-af05-10f2-ae17acbebe52, 'name': SearchDatastore_Task, 'duration_secs': 0.022293} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.445101] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cff3c147-7e32-4c61-8b00-3376b656ba30 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.450772] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 837.450772] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d08e47-de9f-6d2a-61b3-1cd85b88bd66" [ 837.450772] env[61867]: _type = "Task" [ 837.450772] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.465126] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d08e47-de9f-6d2a-61b3-1cd85b88bd66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.559036] env[61867]: DEBUG nova.scheduler.client.report [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 837.640016] env[61867]: DEBUG oslo_vmware.api [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276586, 'name': PowerOnVM_Task, 'duration_secs': 0.869193} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.642733] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 837.643238] env[61867]: INFO nova.compute.manager [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Took 9.29 seconds to spawn the instance on the hypervisor. [ 837.643537] env[61867]: DEBUG nova.compute.manager [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.644460] env[61867]: DEBUG nova.network.neutron [req-bfcdb9a4-ffc4-4d08-b259-87d3cc695bf6 req-868e0028-8250-4e53-8ec5-43c42fbbd30f service nova] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Updated VIF entry in instance network info cache for port eaffa9e8-2616-4b49-9668-c8cc7ef884c0. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 837.644914] env[61867]: DEBUG nova.network.neutron [req-bfcdb9a4-ffc4-4d08-b259-87d3cc695bf6 req-868e0028-8250-4e53-8ec5-43c42fbbd30f service nova] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Updating instance_info_cache with network_info: [{"id": "eaffa9e8-2616-4b49-9668-c8cc7ef884c0", "address": "fa:16:3e:ec:73:fc", "network": {"id": "ee84973a-7c75-4c60-8bb4-8da65e165668", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-732335052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d177558d263c4c8295ef9406bd4aee99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fa01fe1a-83b6-4c10-af75-00ddb17f9bbf", "external-id": "nsx-vlan-transportzone-431", "segmentation_id": 431, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeaffa9e8-26", "ovs_interfaceid": "eaffa9e8-2616-4b49-9668-c8cc7ef884c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.646747] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1da115-c1a6-465e-b323-000a183592d2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.690292] env[61867]: DEBUG nova.compute.manager [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 837.724771] env[61867]: DEBUG nova.network.neutron [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Updating instance_info_cache with network_info: [{"id": "af89c243-c119-47c9-9eda-60c9be2b8c51", "address": "fa:16:3e:4d:1f:76", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf89c243-c1", "ovs_interfaceid": "af89c243-c119-47c9-9eda-60c9be2b8c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.732018] env[61867]: DEBUG nova.virt.hardware [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 837.732018] env[61867]: DEBUG nova.virt.hardware [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 837.732018] env[61867]: DEBUG nova.virt.hardware [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 837.732224] env[61867]: DEBUG nova.virt.hardware [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 837.732224] env[61867]: DEBUG nova.virt.hardware [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 837.732224] env[61867]: DEBUG nova.virt.hardware [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 837.732422] env[61867]: DEBUG nova.virt.hardware [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 837.732752] env[61867]: DEBUG nova.virt.hardware [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 837.733169] env[61867]: DEBUG nova.virt.hardware [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 837.733418] env[61867]: DEBUG nova.virt.hardware [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 837.733660] env[61867]: DEBUG nova.virt.hardware [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 837.734854] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af6294e7-0c10-4f15-a175-952d9eb55b5b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.749022] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f81ab7-cd4d-4b02-8a1d-05d1dd6e5053 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.787818] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': task-1276589, 'name': ReconfigVM_Task, 'duration_secs': 0.34341} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.788229] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Reconfigured VM instance instance-00000047 to attach disk [datastore2] ad699b24-d01d-4d7a-815f-c6b10286012d/ad699b24-d01d-4d7a-815f-c6b10286012d.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.788951] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6505b695-8ff7-483a-9a22-8e80af0ad213 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.795417] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Waiting for the task: (returnval){ [ 837.795417] env[61867]: value = "task-1276590" [ 837.795417] env[61867]: _type = "Task" [ 837.795417] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.803695] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': task-1276590, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.961383] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d08e47-de9f-6d2a-61b3-1cd85b88bd66, 'name': SearchDatastore_Task, 'duration_secs': 0.013716} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.962077] env[61867]: DEBUG oslo_concurrency.lockutils [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.962077] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 8a83f4f2-58eb-473e-9b1e-32ce633554f9/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk. {{(pid=61867) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 837.962317] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3e739dc7-3ef4-4b3f-aa37-b480fc7c5414 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.969796] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 837.969796] env[61867]: value = "task-1276591" [ 837.969796] env[61867]: _type = "Task" [ 837.969796] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.978454] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276591, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.063878] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.408s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.063978] env[61867]: INFO nova.compute.manager [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Migrating [ 838.064194] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.064348] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.066931] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.436s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.066931] env[61867]: DEBUG nova.objects.instance [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Lazy-loading 'resources' on Instance uuid 4653ce96-257e-46d8-8c3d-85c03380213c {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 838.142635] env[61867]: DEBUG nova.compute.manager [req-f8787eb6-dbe7-4cc0-aa1e-fbec1f9ba01c req-b7e773e1-51b5-4e5a-bae3-b14f770a7781 service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Received event network-changed-af89c243-c119-47c9-9eda-60c9be2b8c51 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.142986] env[61867]: DEBUG nova.compute.manager [req-f8787eb6-dbe7-4cc0-aa1e-fbec1f9ba01c req-b7e773e1-51b5-4e5a-bae3-b14f770a7781 service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Refreshing instance network info cache due to event network-changed-af89c243-c119-47c9-9eda-60c9be2b8c51. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 838.142986] env[61867]: DEBUG oslo_concurrency.lockutils [req-f8787eb6-dbe7-4cc0-aa1e-fbec1f9ba01c req-b7e773e1-51b5-4e5a-bae3-b14f770a7781 service nova] Acquiring lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.150896] env[61867]: DEBUG oslo_concurrency.lockutils [req-bfcdb9a4-ffc4-4d08-b259-87d3cc695bf6 req-868e0028-8250-4e53-8ec5-43c42fbbd30f service nova] Releasing lock "refresh_cache-6dda389d-a8c5-4e0e-87a5-4065e24c034e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.164612] env[61867]: INFO nova.compute.manager [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Took 35.05 seconds to build instance. [ 838.227278] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.227673] env[61867]: DEBUG nova.compute.manager [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Instance network_info: |[{"id": "af89c243-c119-47c9-9eda-60c9be2b8c51", "address": "fa:16:3e:4d:1f:76", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf89c243-c1", "ovs_interfaceid": "af89c243-c119-47c9-9eda-60c9be2b8c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 838.227992] env[61867]: DEBUG oslo_concurrency.lockutils [req-f8787eb6-dbe7-4cc0-aa1e-fbec1f9ba01c req-b7e773e1-51b5-4e5a-bae3-b14f770a7781 service nova] Acquired lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.228196] env[61867]: DEBUG nova.network.neutron [req-f8787eb6-dbe7-4cc0-aa1e-fbec1f9ba01c req-b7e773e1-51b5-4e5a-bae3-b14f770a7781 service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Refreshing network info cache for port af89c243-c119-47c9-9eda-60c9be2b8c51 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 838.229426] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:1f:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '11da2092-76f7-447e-babb-8fc14ad39a71', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'af89c243-c119-47c9-9eda-60c9be2b8c51', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 838.237838] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Creating folder: Project (86ced20eaf4740e298dc6f8ca5550c09). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 838.241615] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3997f2f0-d895-465d-9ddf-b36036558a28 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.252478] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Created folder: Project (86ced20eaf4740e298dc6f8ca5550c09) in parent group-v274258. [ 838.252689] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Creating folder: Instances. Parent ref: group-v274342. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 838.252933] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-569d1a88-a668-4ec5-8bee-bd249f2c311e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.261106] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Created folder: Instances in parent group-v274342. [ 838.261350] env[61867]: DEBUG oslo.service.loopingcall [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 838.261567] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 838.261865] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fae21c0b-c5ca-4b5a-888f-5cb931f9afa8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.283163] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 838.283163] env[61867]: value = "task-1276594" [ 838.283163] env[61867]: _type = "Task" [ 838.283163] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.292247] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276594, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.303783] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': task-1276590, 'name': Rename_Task, 'duration_secs': 0.1417} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.304057] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 838.304297] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-016fa1ea-7d11-4f01-af4e-ddd32abc1664 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.313342] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Waiting for the task: (returnval){ [ 838.313342] env[61867]: value = "task-1276595" [ 838.313342] env[61867]: _type = "Task" [ 838.313342] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.318903] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': task-1276595, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.356750] env[61867]: DEBUG nova.network.neutron [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Successfully updated port: cf5eac56-8386-49e1-a2b3-d19e3d8297d6 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 838.481584] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276591, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.568491] env[61867]: INFO nova.compute.rpcapi [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 838.569067] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.579153] env[61867]: INFO nova.compute.manager [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Rebuilding instance [ 838.656887] env[61867]: DEBUG nova.compute.manager [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 838.658082] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ade33e-4e03-45ac-9d33-dbdeebb5e128 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.667260] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fb391f87-69c0-4659-889c-84f93e48731c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "9aab8852-addb-49e6-a59b-fa9bffc7733b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.753s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.709098] env[61867]: DEBUG nova.network.neutron [req-f8787eb6-dbe7-4cc0-aa1e-fbec1f9ba01c req-b7e773e1-51b5-4e5a-bae3-b14f770a7781 service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Updated VIF entry in instance network info cache for port af89c243-c119-47c9-9eda-60c9be2b8c51. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 838.709463] env[61867]: DEBUG nova.network.neutron [req-f8787eb6-dbe7-4cc0-aa1e-fbec1f9ba01c req-b7e773e1-51b5-4e5a-bae3-b14f770a7781 service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Updating instance_info_cache with network_info: [{"id": "af89c243-c119-47c9-9eda-60c9be2b8c51", "address": "fa:16:3e:4d:1f:76", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf89c243-c1", "ovs_interfaceid": "af89c243-c119-47c9-9eda-60c9be2b8c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.794059] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276594, 'name': CreateVM_Task, 'duration_secs': 0.366106} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.794755] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 838.795405] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.795627] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.796197] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 838.796507] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6eb3dfd-4db8-4e5b-8944-1186d13fe6fe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.800876] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 838.800876] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e7bc80-7fc7-75c1-9e1d-baba0cecf5bb" [ 838.800876] env[61867]: _type = "Task" [ 838.800876] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.809126] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e7bc80-7fc7-75c1-9e1d-baba0cecf5bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.819208] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': task-1276595, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.860268] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Acquiring lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.860712] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Acquired lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.860890] env[61867]: DEBUG nova.network.neutron [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 838.895459] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fccb48e-0e5f-4f10-8041-72756d76c2e9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.900464] env[61867]: DEBUG nova.compute.manager [req-01c3b875-9102-4c14-b486-bd986065eb6e req-919c5b93-c4e7-479e-8413-ab41f638de1d service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Received event network-vif-plugged-cf5eac56-8386-49e1-a2b3-d19e3d8297d6 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.901160] env[61867]: DEBUG oslo_concurrency.lockutils [req-01c3b875-9102-4c14-b486-bd986065eb6e req-919c5b93-c4e7-479e-8413-ab41f638de1d service nova] Acquiring lock "410a2819-ea27-4613-9f2c-279f9e82a4b3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.901434] env[61867]: DEBUG oslo_concurrency.lockutils [req-01c3b875-9102-4c14-b486-bd986065eb6e req-919c5b93-c4e7-479e-8413-ab41f638de1d service nova] Lock "410a2819-ea27-4613-9f2c-279f9e82a4b3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.901655] env[61867]: DEBUG oslo_concurrency.lockutils [req-01c3b875-9102-4c14-b486-bd986065eb6e req-919c5b93-c4e7-479e-8413-ab41f638de1d service nova] Lock "410a2819-ea27-4613-9f2c-279f9e82a4b3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.901872] env[61867]: DEBUG nova.compute.manager [req-01c3b875-9102-4c14-b486-bd986065eb6e req-919c5b93-c4e7-479e-8413-ab41f638de1d service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] No waiting events found dispatching network-vif-plugged-cf5eac56-8386-49e1-a2b3-d19e3d8297d6 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 838.902105] env[61867]: WARNING nova.compute.manager [req-01c3b875-9102-4c14-b486-bd986065eb6e req-919c5b93-c4e7-479e-8413-ab41f638de1d service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Received unexpected event network-vif-plugged-cf5eac56-8386-49e1-a2b3-d19e3d8297d6 for instance with vm_state building and task_state spawning. [ 838.902358] env[61867]: DEBUG nova.compute.manager [req-01c3b875-9102-4c14-b486-bd986065eb6e req-919c5b93-c4e7-479e-8413-ab41f638de1d service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Received event network-changed-cf5eac56-8386-49e1-a2b3-d19e3d8297d6 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.902576] env[61867]: DEBUG nova.compute.manager [req-01c3b875-9102-4c14-b486-bd986065eb6e req-919c5b93-c4e7-479e-8413-ab41f638de1d service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Refreshing instance network info cache due to event network-changed-cf5eac56-8386-49e1-a2b3-d19e3d8297d6. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 838.902793] env[61867]: DEBUG oslo_concurrency.lockutils [req-01c3b875-9102-4c14-b486-bd986065eb6e req-919c5b93-c4e7-479e-8413-ab41f638de1d service nova] Acquiring lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.907948] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcffd51a-c73d-4f4b-bd47-9065c3989dfe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.941407] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0f47a3a-a9f8-445d-91e4-5d5378f9f9f9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.949148] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2454e3c5-b222-4a98-bdaa-035e74a0f1dc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.962275] env[61867]: DEBUG nova.compute.provider_tree [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.979492] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276591, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.772016} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.979800] env[61867]: INFO nova.virt.vmwareapi.ds_util [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 8a83f4f2-58eb-473e-9b1e-32ce633554f9/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk. [ 838.980629] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9303073d-7147-4374-9843-3012d905920c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.007023] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 8a83f4f2-58eb-473e-9b1e-32ce633554f9/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 839.007322] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f695514b-f436-4665-944f-f233fa914304 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.025932] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 839.025932] env[61867]: value = "task-1276596" [ 839.025932] env[61867]: _type = "Task" [ 839.025932] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.033879] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276596, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.094874] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.095093] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquired lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.095279] env[61867]: DEBUG nova.network.neutron [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 839.175277] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 839.175600] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-48cb51e9-3e51-430b-990a-f6628b628a9e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.184956] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 839.184956] env[61867]: value = "task-1276597" [ 839.184956] env[61867]: _type = "Task" [ 839.184956] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.194235] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276597, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.212963] env[61867]: DEBUG oslo_concurrency.lockutils [req-f8787eb6-dbe7-4cc0-aa1e-fbec1f9ba01c req-b7e773e1-51b5-4e5a-bae3-b14f770a7781 service nova] Releasing lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.312685] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e7bc80-7fc7-75c1-9e1d-baba0cecf5bb, 'name': SearchDatastore_Task, 'duration_secs': 0.023114} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.315710] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.315914] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.316157] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.316304] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.316475] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.316728] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-31eafec0-e982-481a-8a2a-10ea83aa8206 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.323142] env[61867]: DEBUG oslo_vmware.api [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': task-1276595, 'name': PowerOnVM_Task, 'duration_secs': 0.573589} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.323377] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 839.323579] env[61867]: INFO nova.compute.manager [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Took 8.53 seconds to spawn the instance on the hypervisor. [ 839.323820] env[61867]: DEBUG nova.compute.manager [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.324547] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72f16d56-070c-471e-a7d3-c3c428a067be {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.328868] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.329061] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 839.330862] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cecd670f-34ac-42ff-8878-fc57c9dd08b9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.337947] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 839.337947] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520a0ae8-4dda-e79f-8d04-b55a099c50f1" [ 839.337947] env[61867]: _type = "Task" [ 839.337947] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.344915] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520a0ae8-4dda-e79f-8d04-b55a099c50f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.392934] env[61867]: DEBUG nova.network.neutron [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 839.465561] env[61867]: DEBUG nova.scheduler.client.report [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 839.537528] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276596, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.578087] env[61867]: DEBUG nova.network.neutron [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Updating instance_info_cache with network_info: [{"id": "cf5eac56-8386-49e1-a2b3-d19e3d8297d6", "address": "fa:16:3e:a9:7a:17", "network": {"id": "06608a13-6187-41bb-926a-1fbda4f627a0", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-618264839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e54d4ce9d6c24e8ab40cb77886aac634", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf5eac56-83", "ovs_interfaceid": "cf5eac56-8386-49e1-a2b3-d19e3d8297d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.695449] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276597, 'name': PowerOffVM_Task, 'duration_secs': 0.473923} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.695676] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 839.695897] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 839.696667] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7be62a6-8a33-4858-913f-6956658c6bc6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.705987] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 839.706246] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8399305e-ce41-4243-ace5-1a3d6a0759d5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.803299] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 839.803547] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 839.803737] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleting the datastore file [datastore2] 9aab8852-addb-49e6-a59b-fa9bffc7733b {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 839.804046] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e0bf216-de5e-4702-9b1e-5582125e93f8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.810867] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 839.810867] env[61867]: value = "task-1276599" [ 839.810867] env[61867]: _type = "Task" [ 839.810867] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.819423] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276599, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.848838] env[61867]: INFO nova.compute.manager [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Took 34.33 seconds to build instance. [ 839.857304] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520a0ae8-4dda-e79f-8d04-b55a099c50f1, 'name': SearchDatastore_Task, 'duration_secs': 0.033397} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.862218] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fc9f58a-490c-4c52-94ed-ca623cfb48b1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.870948] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 839.870948] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52a8189b-ec2a-e7fe-3689-324038c5fc4d" [ 839.870948] env[61867]: _type = "Task" [ 839.870948] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.881224] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52a8189b-ec2a-e7fe-3689-324038c5fc4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.970764] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.905s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.973533] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.357s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.975449] env[61867]: INFO nova.compute.claims [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 839.995307] env[61867]: INFO nova.scheduler.client.report [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Deleted allocations for instance 4653ce96-257e-46d8-8c3d-85c03380213c [ 840.040129] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276596, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.081254] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Releasing lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.081597] env[61867]: DEBUG nova.compute.manager [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Instance network_info: |[{"id": "cf5eac56-8386-49e1-a2b3-d19e3d8297d6", "address": "fa:16:3e:a9:7a:17", "network": {"id": "06608a13-6187-41bb-926a-1fbda4f627a0", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-618264839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e54d4ce9d6c24e8ab40cb77886aac634", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf5eac56-83", "ovs_interfaceid": "cf5eac56-8386-49e1-a2b3-d19e3d8297d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 840.082025] env[61867]: DEBUG oslo_concurrency.lockutils [req-01c3b875-9102-4c14-b486-bd986065eb6e req-919c5b93-c4e7-479e-8413-ab41f638de1d service nova] Acquired lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.082238] env[61867]: DEBUG nova.network.neutron [req-01c3b875-9102-4c14-b486-bd986065eb6e req-919c5b93-c4e7-479e-8413-ab41f638de1d service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Refreshing network info cache for port cf5eac56-8386-49e1-a2b3-d19e3d8297d6 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 840.087233] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:7a:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f856fca-9fb5-41ea-a057-ac4193bd323d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cf5eac56-8386-49e1-a2b3-d19e3d8297d6', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 840.097312] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Creating folder: Project (e54d4ce9d6c24e8ab40cb77886aac634). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 840.098446] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5b432084-fc8c-432a-ae9d-92b600e77fdf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.109602] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Created folder: Project (e54d4ce9d6c24e8ab40cb77886aac634) in parent group-v274258. [ 840.109818] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Creating folder: Instances. Parent ref: group-v274345. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 840.110155] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c9a8cef2-2aa5-4646-a126-c6604e18df5b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.119217] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Created folder: Instances in parent group-v274345. [ 840.119456] env[61867]: DEBUG oslo.service.loopingcall [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.119649] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 840.119857] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0e9e31a3-40dd-4b5f-b669-12aad1d09b34 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.135310] env[61867]: DEBUG nova.network.neutron [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Updating instance_info_cache with network_info: [{"id": "79dddbac-ec4f-4cd9-8436-a3600957e598", "address": "fa:16:3e:b8:10:25", "network": {"id": "2d4ab52c-144f-45a1-9d4f-afd85fc30404", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.63", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "84601e35c8e8487cb78fc16a2536a4c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d47d5e1d-e66d-4f2c-83e6-d5e78c2b767d", "external-id": "nsx-vlan-transportzone-109", "segmentation_id": 109, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79dddbac-ec", "ovs_interfaceid": "79dddbac-ec4f-4cd9-8436-a3600957e598", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.144030] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 840.144030] env[61867]: value = "task-1276602" [ 840.144030] env[61867]: _type = "Task" [ 840.144030] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.152876] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276602, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.321600] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276599, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.242504} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.321886] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.322174] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 840.322267] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 840.351856] env[61867]: DEBUG oslo_concurrency.lockutils [None req-266eaa81-f97e-43fc-8b07-143780ea08c5 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Lock "ad699b24-d01d-4d7a-815f-c6b10286012d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.579s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.366266] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Acquiring lock "ad699b24-d01d-4d7a-815f-c6b10286012d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.366560] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Lock "ad699b24-d01d-4d7a-815f-c6b10286012d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.366792] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Acquiring lock "ad699b24-d01d-4d7a-815f-c6b10286012d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.367015] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Lock "ad699b24-d01d-4d7a-815f-c6b10286012d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.367240] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Lock "ad699b24-d01d-4d7a-815f-c6b10286012d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.369441] env[61867]: INFO nova.compute.manager [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Terminating instance [ 840.371324] env[61867]: DEBUG nova.compute.manager [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 840.371519] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 840.372414] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb56798-09db-487d-a1ae-d7e019ca1eec {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.384201] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52a8189b-ec2a-e7fe-3689-324038c5fc4d, 'name': SearchDatastore_Task, 'duration_secs': 0.025489} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.386325] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.386596] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] ca2f6d18-f773-4875-83de-2f2be912c2f8/ca2f6d18-f773-4875-83de-2f2be912c2f8.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 840.386898] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 840.387265] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5f15a659-5e94-4d34-88f5-d2684402e29f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.388983] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b17e2e8b-d0e6-4c93-a48d-30f781f6fc4d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.395171] env[61867]: DEBUG oslo_vmware.api [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Waiting for the task: (returnval){ [ 840.395171] env[61867]: value = "task-1276604" [ 840.395171] env[61867]: _type = "Task" [ 840.395171] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.396348] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 840.396348] env[61867]: value = "task-1276603" [ 840.396348] env[61867]: _type = "Task" [ 840.396348] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.408482] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276603, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.503416] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f10d6ec2-1f9d-482f-9298-43bec5884e51 tempest-ServerPasswordTestJSON-2051850044 tempest-ServerPasswordTestJSON-2051850044-project-member] Lock "4653ce96-257e-46d8-8c3d-85c03380213c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.256s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.538275] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276596, 'name': ReconfigVM_Task, 'duration_secs': 1.032916} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.538611] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 8a83f4f2-58eb-473e-9b1e-32ce633554f9/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 840.539541] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e7817cc-7412-41b6-ac4b-e0478afdedda {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.565903] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56eb5093-7255-49f0-b004-578c78f35a39 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.580447] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 840.580447] env[61867]: value = "task-1276605" [ 840.580447] env[61867]: _type = "Task" [ 840.580447] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.590103] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276605, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.638960] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Releasing lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.653266] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276602, 'name': CreateVM_Task, 'duration_secs': 0.418317} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.654238] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 840.654932] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.655115] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.655463] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 840.656015] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46479b3a-c479-4f48-bb3c-2105beae2da3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.660927] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 840.660927] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f59f91-cb87-603f-2417-36a28dda8bc8" [ 840.660927] env[61867]: _type = "Task" [ 840.660927] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.671388] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f59f91-cb87-603f-2417-36a28dda8bc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.913124] env[61867]: DEBUG oslo_vmware.api [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': task-1276604, 'name': PowerOffVM_Task, 'duration_secs': 0.191604} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.913349] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276603, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493191} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.914099] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 840.914099] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 840.914099] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] ca2f6d18-f773-4875-83de-2f2be912c2f8/ca2f6d18-f773-4875-83de-2f2be912c2f8.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 840.914240] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.914428] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b3c620f4-a5f0-4149-8bb9-33d8f8b95c74 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.915943] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0387764d-6448-40f4-b2ac-911fff6d6d47 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.923742] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 840.923742] env[61867]: value = "task-1276607" [ 840.923742] env[61867]: _type = "Task" [ 840.923742] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.924694] env[61867]: DEBUG nova.network.neutron [req-01c3b875-9102-4c14-b486-bd986065eb6e req-919c5b93-c4e7-479e-8413-ab41f638de1d service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Updated VIF entry in instance network info cache for port cf5eac56-8386-49e1-a2b3-d19e3d8297d6. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 840.925038] env[61867]: DEBUG nova.network.neutron [req-01c3b875-9102-4c14-b486-bd986065eb6e req-919c5b93-c4e7-479e-8413-ab41f638de1d service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Updating instance_info_cache with network_info: [{"id": "cf5eac56-8386-49e1-a2b3-d19e3d8297d6", "address": "fa:16:3e:a9:7a:17", "network": {"id": "06608a13-6187-41bb-926a-1fbda4f627a0", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-618264839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e54d4ce9d6c24e8ab40cb77886aac634", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf5eac56-83", "ovs_interfaceid": "cf5eac56-8386-49e1-a2b3-d19e3d8297d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.935868] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276607, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.006710] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 841.007086] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 841.007391] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Deleting the datastore file [datastore2] ad699b24-d01d-4d7a-815f-c6b10286012d {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 841.011126] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc59f0c7-d73e-4dad-8836-c6c685619608 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.018569] env[61867]: DEBUG oslo_vmware.api [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Waiting for the task: (returnval){ [ 841.018569] env[61867]: value = "task-1276608" [ 841.018569] env[61867]: _type = "Task" [ 841.018569] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.029663] env[61867]: DEBUG oslo_vmware.api [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': task-1276608, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.089969] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276605, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.171458] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f59f91-cb87-603f-2417-36a28dda8bc8, 'name': SearchDatastore_Task, 'duration_secs': 0.050221} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.174149] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.174401] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 841.174634] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.174832] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.174964] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 841.175427] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-15fdeafe-8576-4f4e-a578-61a6e57e278a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.183995] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 841.184070] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 841.187259] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06ee289c-61de-452b-8436-2fe2f17dd94a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.192662] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 841.192662] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5244b19c-8e26-07dd-fca6-1fd4df1a1865" [ 841.192662] env[61867]: _type = "Task" [ 841.192662] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.201636] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5244b19c-8e26-07dd-fca6-1fd4df1a1865, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.249816] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aba5c94-ca8c-42b5-aefb-9cca4494b7d6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.257600] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d4eedc-0ff4-420d-87cd-df40982c38cf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.294371] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef9be73-fd89-42be-bae5-4aa7ff8c8b3b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.302563] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c572f4f9-152d-47af-a348-53db4578b85e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.316791] env[61867]: DEBUG nova.compute.provider_tree [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 841.360683] env[61867]: DEBUG nova.virt.hardware [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 841.361366] env[61867]: DEBUG nova.virt.hardware [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 841.361717] env[61867]: DEBUG nova.virt.hardware [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 841.365017] env[61867]: DEBUG nova.virt.hardware [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 841.365017] env[61867]: DEBUG nova.virt.hardware [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 841.365017] env[61867]: DEBUG nova.virt.hardware [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 841.365017] env[61867]: DEBUG nova.virt.hardware [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 841.365017] env[61867]: DEBUG nova.virt.hardware [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 841.365222] env[61867]: DEBUG nova.virt.hardware [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 841.365222] env[61867]: DEBUG nova.virt.hardware [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 841.365222] env[61867]: DEBUG nova.virt.hardware [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 841.365222] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa63f91-3f1b-43b0-ae89-0ea95df32d67 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.374597] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f36e5369-788f-4ea8-a379-d7c68e3d1bd3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.396451] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:5e:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f2020f39-42c4-4481-85c5-aaf03854b459', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '104495f6-b976-4f53-9959-d5193f833ae8', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 841.411133] env[61867]: DEBUG oslo.service.loopingcall [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.412336] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 841.412829] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c00dcfab-d96f-457e-b4b7-d11c7370c092 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.439587] env[61867]: DEBUG oslo_concurrency.lockutils [req-01c3b875-9102-4c14-b486-bd986065eb6e req-919c5b93-c4e7-479e-8413-ab41f638de1d service nova] Releasing lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.449181] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276607, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.450366] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 841.450366] env[61867]: value = "task-1276609" [ 841.450366] env[61867]: _type = "Task" [ 841.450366] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.457516] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276609, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.528169] env[61867]: DEBUG oslo_vmware.api [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Task: {'id': task-1276608, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.208131} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.528344] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 841.528538] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 841.528719] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 841.528895] env[61867]: INFO nova.compute.manager [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 841.529155] env[61867]: DEBUG oslo.service.loopingcall [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.529352] env[61867]: DEBUG nova.compute.manager [-] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 841.529446] env[61867]: DEBUG nova.network.neutron [-] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 841.593260] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276605, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.703232] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5244b19c-8e26-07dd-fca6-1fd4df1a1865, 'name': SearchDatastore_Task, 'duration_secs': 0.008186} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.704134] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea9ac66e-2470-438a-833a-b593fe8cefa2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.709521] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 841.709521] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528f92c6-93e6-bf0e-f288-558d3b4184de" [ 841.709521] env[61867]: _type = "Task" [ 841.709521] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.717488] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528f92c6-93e6-bf0e-f288-558d3b4184de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.837550] env[61867]: ERROR nova.scheduler.client.report [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [req-d1c71fa8-0e66-4de5-9b6b-ef5f2c01ba0d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 25720271-a549-4916-abe3-e5ed9b765889. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-d1c71fa8-0e66-4de5-9b6b-ef5f2c01ba0d"}]} [ 841.857225] env[61867]: DEBUG nova.scheduler.client.report [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Refreshing inventories for resource provider 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 841.879162] env[61867]: DEBUG nova.scheduler.client.report [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Updating ProviderTree inventory for provider 25720271-a549-4916-abe3-e5ed9b765889 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 841.879432] env[61867]: DEBUG nova.compute.provider_tree [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 841.902608] env[61867]: DEBUG nova.scheduler.client.report [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Refreshing aggregate associations for resource provider 25720271-a549-4916-abe3-e5ed9b765889, aggregates: None {{(pid=61867) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 841.930388] env[61867]: DEBUG nova.scheduler.client.report [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Refreshing trait associations for resource provider 25720271-a549-4916-abe3-e5ed9b765889, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61867) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 841.949451] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276607, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.959308] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276609, 'name': CreateVM_Task} progress is 25%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.097877] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276605, 'name': ReconfigVM_Task, 'duration_secs': 1.183586} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.098252] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.098808] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2670b31-44d4-49c3-b60e-a6a81402ca69 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.105623] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 842.105623] env[61867]: value = "task-1276610" [ 842.105623] env[61867]: _type = "Task" [ 842.105623] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.118114] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276610, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.143235] env[61867]: DEBUG nova.compute.manager [req-521ae850-0471-4283-9899-7c5bf4a475ea req-20656ab3-05bc-45ca-935b-944966ba496d service nova] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Received event network-vif-deleted-880e13b8-9318-4197-8bab-3bbc6ddcf952 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.143492] env[61867]: INFO nova.compute.manager [req-521ae850-0471-4283-9899-7c5bf4a475ea req-20656ab3-05bc-45ca-935b-944966ba496d service nova] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Neutron deleted interface 880e13b8-9318-4197-8bab-3bbc6ddcf952; detaching it from the instance and deleting it from the info cache [ 842.143638] env[61867]: DEBUG nova.network.neutron [req-521ae850-0471-4283-9899-7c5bf4a475ea req-20656ab3-05bc-45ca-935b-944966ba496d service nova] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.163831] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-185b25ab-9538-473d-a6c0-f5b7d3b42fa0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.184431] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Updating instance '81e70c36-04ea-450c-9383-53ef069d1c46' progress to 0 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 842.222815] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528f92c6-93e6-bf0e-f288-558d3b4184de, 'name': SearchDatastore_Task, 'duration_secs': 0.054424} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.223133] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.223400] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 410a2819-ea27-4613-9f2c-279f9e82a4b3/410a2819-ea27-4613-9f2c-279f9e82a4b3.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 842.223698] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b9784957-e024-4c3a-bca8-afa6f3bf2c77 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.230191] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 842.230191] env[61867]: value = "task-1276611" [ 842.230191] env[61867]: _type = "Task" [ 842.230191] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.241228] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276611, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.267499] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec35963-53a5-41c3-8cfa-8b830da9461b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.275422] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db447cac-54d9-4fb5-82ef-fcbd5ccc00d9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.305752] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98d337e7-60ea-405b-a565-4875eab98594 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.313474] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cdc577f-e80a-4203-be76-6d8f82ca2fc1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.327284] env[61867]: DEBUG nova.compute.provider_tree [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 842.450924] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276607, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.074949} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.452128] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 842.455960] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-797eade8-1041-48ff-9acb-bcbd0ca9c41b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.463370] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276609, 'name': CreateVM_Task, 'duration_secs': 0.799453} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.473027] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 842.483450] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] ca2f6d18-f773-4875-83de-2f2be912c2f8/ca2f6d18-f773-4875-83de-2f2be912c2f8.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 842.484648] env[61867]: DEBUG oslo_concurrency.lockutils [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.484924] env[61867]: DEBUG oslo_concurrency.lockutils [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.485303] env[61867]: DEBUG oslo_concurrency.lockutils [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 842.485610] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-efaa03c0-45dc-4fe9-88db-35b0ccbd6a35 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.500661] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b60e342-5770-40bb-bbcd-5abb50758fc7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.508708] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 842.508708] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5295ba70-0aa6-4308-e734-be58f31b18f4" [ 842.508708] env[61867]: _type = "Task" [ 842.508708] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.512788] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 842.512788] env[61867]: value = "task-1276612" [ 842.512788] env[61867]: _type = "Task" [ 842.512788] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.520096] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5295ba70-0aa6-4308-e734-be58f31b18f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.525086] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276612, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.599051] env[61867]: DEBUG nova.network.neutron [-] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.617173] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276610, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.648300] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fe57a336-6d3d-459c-8323-ab2f1a796681 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.657455] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f5e38f6-9418-45da-9f9f-26763b039310 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.687276] env[61867]: DEBUG nova.compute.manager [req-521ae850-0471-4283-9899-7c5bf4a475ea req-20656ab3-05bc-45ca-935b-944966ba496d service nova] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Detach interface failed, port_id=880e13b8-9318-4197-8bab-3bbc6ddcf952, reason: Instance ad699b24-d01d-4d7a-815f-c6b10286012d could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 842.692940] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 842.695941] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b1c55697-4c1e-4e5d-8439-402f17392c75 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.701048] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 842.701048] env[61867]: value = "task-1276613" [ 842.701048] env[61867]: _type = "Task" [ 842.701048] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.709203] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276613, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.740763] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276611, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.860542] env[61867]: DEBUG nova.scheduler.client.report [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Updated inventory for provider 25720271-a549-4916-abe3-e5ed9b765889 with generation 88 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 842.860958] env[61867]: DEBUG nova.compute.provider_tree [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Updating resource provider 25720271-a549-4916-abe3-e5ed9b765889 generation from 88 to 89 during operation: update_inventory {{(pid=61867) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 842.861228] env[61867]: DEBUG nova.compute.provider_tree [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 843.023614] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5295ba70-0aa6-4308-e734-be58f31b18f4, 'name': SearchDatastore_Task, 'duration_secs': 0.00981} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.024493] env[61867]: DEBUG oslo_concurrency.lockutils [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.024771] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 843.025021] env[61867]: DEBUG oslo_concurrency.lockutils [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.025191] env[61867]: DEBUG oslo_concurrency.lockutils [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.025396] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.028975] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-661789a2-5a5a-4bb4-9bc2-136b863fca8b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.031301] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.040142] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.040396] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 843.041320] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4277f590-2e90-4103-98c7-d30d188befbf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.047161] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 843.047161] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5239eb11-d19d-8a5d-83dd-5dbeeb60c42d" [ 843.047161] env[61867]: _type = "Task" [ 843.047161] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.054949] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5239eb11-d19d-8a5d-83dd-5dbeeb60c42d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.104409] env[61867]: INFO nova.compute.manager [-] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Took 1.57 seconds to deallocate network for instance. [ 843.118719] env[61867]: DEBUG oslo_vmware.api [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276610, 'name': PowerOnVM_Task, 'duration_secs': 0.654823} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.120186] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.122583] env[61867]: DEBUG nova.compute.manager [None req-db4e414b-be8b-4985-bd09-9f89d246980e tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.123560] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-054a1047-ad93-4d26-88a4-b6025087578f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.214910] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276613, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.241904] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276611, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.366983] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.393s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.367695] env[61867]: DEBUG nova.compute.manager [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 843.370830] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.456s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.372772] env[61867]: INFO nova.compute.claims [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.523954] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276612, 'name': ReconfigVM_Task, 'duration_secs': 0.577923} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.524275] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Reconfigured VM instance instance-00000048 to attach disk [datastore1] ca2f6d18-f773-4875-83de-2f2be912c2f8/ca2f6d18-f773-4875-83de-2f2be912c2f8.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 843.524927] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f0ad0996-d55b-4491-80f3-ad0f217d4151 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.530466] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 843.530466] env[61867]: value = "task-1276614" [ 843.530466] env[61867]: _type = "Task" [ 843.530466] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.537669] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276614, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.556169] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5239eb11-d19d-8a5d-83dd-5dbeeb60c42d, 'name': SearchDatastore_Task, 'duration_secs': 0.009521} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.556894] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-237af84b-b965-4df8-a4ba-cd8ae65b376b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.561740] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 843.561740] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d33f83-0209-fda2-df66-9fb8e28b081a" [ 843.561740] env[61867]: _type = "Task" [ 843.561740] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.569076] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d33f83-0209-fda2-df66-9fb8e28b081a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.615468] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.711820] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276613, 'name': PowerOffVM_Task, 'duration_secs': 0.523723} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.712139] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 843.712334] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Updating instance '81e70c36-04ea-450c-9383-53ef069d1c46' progress to 17 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 843.744301] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276611, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.183752} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.744610] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 410a2819-ea27-4613-9f2c-279f9e82a4b3/410a2819-ea27-4613-9f2c-279f9e82a4b3.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 843.744831] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 843.745189] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4d133867-d056-4c18-bf19-30ee52b80a3a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.751934] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 843.751934] env[61867]: value = "task-1276615" [ 843.751934] env[61867]: _type = "Task" [ 843.751934] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.761640] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276615, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.877633] env[61867]: DEBUG nova.compute.utils [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 843.881880] env[61867]: DEBUG nova.compute.manager [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 843.882138] env[61867]: DEBUG nova.network.neutron [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 843.919678] env[61867]: DEBUG nova.policy [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc358009ebe6495a881034439d00978f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a29c5e9cfdaa4cc88ed300100a308ffa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 844.041282] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276614, 'name': Rename_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.073439] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d33f83-0209-fda2-df66-9fb8e28b081a, 'name': SearchDatastore_Task, 'duration_secs': 0.009006} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.074108] env[61867]: DEBUG oslo_concurrency.lockutils [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.074108] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 9aab8852-addb-49e6-a59b-fa9bffc7733b/9aab8852-addb-49e6-a59b-fa9bffc7733b.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 844.074314] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-08aa5e0d-2c53-4c4a-8ba6-c33a7985867d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.080808] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 844.080808] env[61867]: value = "task-1276616" [ 844.080808] env[61867]: _type = "Task" [ 844.080808] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.088685] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276616, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.220315] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 844.220698] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 844.220944] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.221216] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 844.221391] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.221593] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 844.221816] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 844.221990] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 844.222178] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 844.222348] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 844.222524] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.227625] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9198b41-648d-4616-b4bf-4df6491fce0b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.244360] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 844.244360] env[61867]: value = "task-1276617" [ 844.244360] env[61867]: _type = "Task" [ 844.244360] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.252766] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276617, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.260133] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276615, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.277093} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.260390] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 844.261220] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07783bcd-a200-480e-96b8-416230e58890 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.294817] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 410a2819-ea27-4613-9f2c-279f9e82a4b3/410a2819-ea27-4613-9f2c-279f9e82a4b3.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 844.295248] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ffcce1ed-bd87-4d54-8c1b-ec0de69d89d2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.324077] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 844.324077] env[61867]: value = "task-1276618" [ 844.324077] env[61867]: _type = "Task" [ 844.324077] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.332439] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276618, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.379047] env[61867]: DEBUG nova.network.neutron [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Successfully created port: 8fe6f5be-9448-4542-a731-fc42b5ab06ca {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 844.384830] env[61867]: DEBUG nova.compute.manager [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 844.544332] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276614, 'name': Rename_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.590423] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276616, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.612718] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a42ebea6-fb00-4668-9ef0-ecdf7b4482ed {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.620214] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b69d2e7-fdd5-4ad8-b9b1-f33d375c1b72 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.652187] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7639d31b-39eb-480a-a584-46da3971bc1d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.659923] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9556ed95-41cb-4618-813b-9b395e4da0f6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.673812] env[61867]: DEBUG nova.compute.provider_tree [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.755694] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276617, 'name': ReconfigVM_Task, 'duration_secs': 0.461014} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.756113] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Updating instance '81e70c36-04ea-450c-9383-53ef069d1c46' progress to 33 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 844.837141] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276618, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.044964] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276614, 'name': Rename_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.091090] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276616, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.923897} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.091409] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 9aab8852-addb-49e6-a59b-fa9bffc7733b/9aab8852-addb-49e6-a59b-fa9bffc7733b.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 845.091630] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.091906] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1d7c5574-8a09-4fb2-8033-8fa441fe761c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.097778] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 845.097778] env[61867]: value = "task-1276619" [ 845.097778] env[61867]: _type = "Task" [ 845.097778] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.105360] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276619, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.176650] env[61867]: DEBUG nova.scheduler.client.report [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.262848] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:47:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='f8967eee-6d82-41b2-98fd-bc50fb8c603b',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-993246556',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 845.263150] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 845.263320] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 845.263507] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 845.263655] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 845.263807] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 845.264254] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 845.264489] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 845.264710] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 845.264933] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 845.265176] env[61867]: DEBUG nova.virt.hardware [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 845.271100] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Reconfiguring VM instance instance-0000003f to detach disk 2000 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 845.271392] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c81533e-32d2-42c6-9dd8-68c900f794bd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.289844] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 845.289844] env[61867]: value = "task-1276620" [ 845.289844] env[61867]: _type = "Task" [ 845.289844] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.297497] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276620, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.336204] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276618, 'name': ReconfigVM_Task, 'duration_secs': 0.792017} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.336354] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 410a2819-ea27-4613-9f2c-279f9e82a4b3/410a2819-ea27-4613-9f2c-279f9e82a4b3.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 845.337216] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fa4b5e99-3d8f-4d69-946d-4b49bfb733d7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.346308] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 845.346308] env[61867]: value = "task-1276621" [ 845.346308] env[61867]: _type = "Task" [ 845.346308] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.353293] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276621, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.399040] env[61867]: DEBUG nova.compute.manager [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 845.423979] env[61867]: DEBUG nova.virt.hardware [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='70ddcb7b1b023e15c3fb4d2b2a1972da',container_format='bare',created_at=2024-10-12T12:48:43Z,direct_url=,disk_format='vmdk',id=420f5356-3ba1-4114-af3a-5be2a65b82cb,min_disk=1,min_ram=0,name='tempest-test-snap-110435987',owner='a29c5e9cfdaa4cc88ed300100a308ffa',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-12T12:49:01Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 845.424281] env[61867]: DEBUG nova.virt.hardware [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 845.424444] env[61867]: DEBUG nova.virt.hardware [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 845.424672] env[61867]: DEBUG nova.virt.hardware [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 845.424840] env[61867]: DEBUG nova.virt.hardware [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 845.424995] env[61867]: DEBUG nova.virt.hardware [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 845.425230] env[61867]: DEBUG nova.virt.hardware [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 845.425395] env[61867]: DEBUG nova.virt.hardware [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 845.425567] env[61867]: DEBUG nova.virt.hardware [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 845.425735] env[61867]: DEBUG nova.virt.hardware [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 845.425914] env[61867]: DEBUG nova.virt.hardware [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 845.426793] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d2bde53-4afb-45f0-9c08-c24606fac640 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.434908] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b547fa-50b7-41c0-87d9-512b7cc67cb9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.544157] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276614, 'name': Rename_Task, 'duration_secs': 1.63418} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.544445] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 845.544693] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fcdb5513-4ef3-4c6a-b7e4-c1a985112ff5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.550805] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 845.550805] env[61867]: value = "task-1276622" [ 845.550805] env[61867]: _type = "Task" [ 845.550805] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.558493] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276622, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.606589] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276619, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084864} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.606881] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 845.607733] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab78de7-a160-4a65-8f9d-7476786b8e6b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.628978] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 9aab8852-addb-49e6-a59b-fa9bffc7733b/9aab8852-addb-49e6-a59b-fa9bffc7733b.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 845.629573] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ceb6e942-b4e3-4287-bd31-01779b90d6f5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.648265] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 845.648265] env[61867]: value = "task-1276623" [ 845.648265] env[61867]: _type = "Task" [ 845.648265] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.655672] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276623, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.681815] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.311s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.682419] env[61867]: DEBUG nova.compute.manager [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 845.684985] env[61867]: DEBUG oslo_concurrency.lockutils [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.945s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.688104] env[61867]: INFO nova.compute.claims [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.795985] env[61867]: DEBUG nova.compute.manager [req-738b5d56-75f2-4b19-9b24-e33cf49c7a31 req-fb64dae9-2a88-4e97-a744-bafa2ea2ce23 service nova] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Received event network-vif-plugged-8fe6f5be-9448-4542-a731-fc42b5ab06ca {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.796236] env[61867]: DEBUG oslo_concurrency.lockutils [req-738b5d56-75f2-4b19-9b24-e33cf49c7a31 req-fb64dae9-2a88-4e97-a744-bafa2ea2ce23 service nova] Acquiring lock "e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.796444] env[61867]: DEBUG oslo_concurrency.lockutils [req-738b5d56-75f2-4b19-9b24-e33cf49c7a31 req-fb64dae9-2a88-4e97-a744-bafa2ea2ce23 service nova] Lock "e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.796613] env[61867]: DEBUG oslo_concurrency.lockutils [req-738b5d56-75f2-4b19-9b24-e33cf49c7a31 req-fb64dae9-2a88-4e97-a744-bafa2ea2ce23 service nova] Lock "e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.796801] env[61867]: DEBUG nova.compute.manager [req-738b5d56-75f2-4b19-9b24-e33cf49c7a31 req-fb64dae9-2a88-4e97-a744-bafa2ea2ce23 service nova] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] No waiting events found dispatching network-vif-plugged-8fe6f5be-9448-4542-a731-fc42b5ab06ca {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 845.796967] env[61867]: WARNING nova.compute.manager [req-738b5d56-75f2-4b19-9b24-e33cf49c7a31 req-fb64dae9-2a88-4e97-a744-bafa2ea2ce23 service nova] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Received unexpected event network-vif-plugged-8fe6f5be-9448-4542-a731-fc42b5ab06ca for instance with vm_state building and task_state spawning. [ 845.803031] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276620, 'name': ReconfigVM_Task, 'duration_secs': 0.195548} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.803292] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Reconfigured VM instance instance-0000003f to detach disk 2000 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 845.804042] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6735f8a-6ff2-47f8-90bc-48873d055d26 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.827723] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 81e70c36-04ea-450c-9383-53ef069d1c46/81e70c36-04ea-450c-9383-53ef069d1c46.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 845.828017] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d81d6921-389a-4f74-ab3f-b0fc447bcd4d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.849226] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 845.849226] env[61867]: value = "task-1276624" [ 845.849226] env[61867]: _type = "Task" [ 845.849226] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.855965] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276621, 'name': Rename_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.861144] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276624, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.993476] env[61867]: INFO nova.compute.manager [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Rescuing [ 845.993746] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "refresh_cache-6e41989e-b8fa-4009-af1e-1ce859b329a1" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.993905] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired lock "refresh_cache-6e41989e-b8fa-4009-af1e-1ce859b329a1" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.994087] env[61867]: DEBUG nova.network.neutron [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 846.062122] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276622, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.159542] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276623, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.193541] env[61867]: DEBUG nova.compute.utils [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 846.198337] env[61867]: DEBUG nova.compute.manager [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 846.198617] env[61867]: DEBUG nova.network.neutron [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 846.259407] env[61867]: DEBUG nova.policy [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a013ca682054d16a6702230a22c180f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c28a8e133444440eb83dc6a848ef591d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 846.357575] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276621, 'name': Rename_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.363197] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276624, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.367870] env[61867]: DEBUG nova.network.neutron [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Successfully updated port: 8fe6f5be-9448-4542-a731-fc42b5ab06ca {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 846.392369] env[61867]: DEBUG nova.compute.manager [req-529a7038-8557-459a-8c0b-e418da5452cb req-2ff29609-5200-4d69-9ed5-b4a7a3ef51c2 service nova] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Received event network-changed-8fe6f5be-9448-4542-a731-fc42b5ab06ca {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.392488] env[61867]: DEBUG nova.compute.manager [req-529a7038-8557-459a-8c0b-e418da5452cb req-2ff29609-5200-4d69-9ed5-b4a7a3ef51c2 service nova] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Refreshing instance network info cache due to event network-changed-8fe6f5be-9448-4542-a731-fc42b5ab06ca. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 846.392985] env[61867]: DEBUG oslo_concurrency.lockutils [req-529a7038-8557-459a-8c0b-e418da5452cb req-2ff29609-5200-4d69-9ed5-b4a7a3ef51c2 service nova] Acquiring lock "refresh_cache-e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.392985] env[61867]: DEBUG oslo_concurrency.lockutils [req-529a7038-8557-459a-8c0b-e418da5452cb req-2ff29609-5200-4d69-9ed5-b4a7a3ef51c2 service nova] Acquired lock "refresh_cache-e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.392985] env[61867]: DEBUG nova.network.neutron [req-529a7038-8557-459a-8c0b-e418da5452cb req-2ff29609-5200-4d69-9ed5-b4a7a3ef51c2 service nova] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Refreshing network info cache for port 8fe6f5be-9448-4542-a731-fc42b5ab06ca {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 846.561958] env[61867]: DEBUG oslo_vmware.api [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276622, 'name': PowerOnVM_Task, 'duration_secs': 0.621064} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.562291] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 846.562443] env[61867]: INFO nova.compute.manager [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Took 11.32 seconds to spawn the instance on the hypervisor. [ 846.562663] env[61867]: DEBUG nova.compute.manager [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 846.565577] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15166238-639d-47cd-bb99-72557d44b4df {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.660892] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276623, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.688140] env[61867]: DEBUG nova.network.neutron [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Successfully created port: 2b465a8c-afc2-4890-a172-d8beb1215f8a {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 846.699370] env[61867]: DEBUG nova.compute.manager [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 846.864742] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276621, 'name': Rename_Task, 'duration_secs': 1.1949} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.865008] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276624, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.868425] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 846.868968] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4cd8f332-7b9d-468a-8b63-fcc32fa75cc9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.871660] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "refresh_cache-e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.877688] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 846.877688] env[61867]: value = "task-1276625" [ 846.877688] env[61867]: _type = "Task" [ 846.877688] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.888046] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276625, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.901124] env[61867]: DEBUG nova.network.neutron [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Updating instance_info_cache with network_info: [{"id": "d8069fac-960f-49cf-a416-036e26060a27", "address": "fa:16:3e:ca:61:fe", "network": {"id": "ab6ec5b9-8193-43c9-a3ce-d63f31e5a001", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1573384141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35aa85105b394369b17ef4ce51a49066", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8069fac-96", "ovs_interfaceid": "d8069fac-960f-49cf-a416-036e26060a27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.953563] env[61867]: DEBUG nova.network.neutron [req-529a7038-8557-459a-8c0b-e418da5452cb req-2ff29609-5200-4d69-9ed5-b4a7a3ef51c2 service nova] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 847.021670] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b2ed19-c5ba-4ce8-b87b-b47b75d2ef34 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.030684] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b2d59f-11de-43f9-bab0-029d8fac6f04 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.084535] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4272d45-4901-4a8b-bc7c-7dd8a6613a85 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.088321] env[61867]: INFO nova.compute.manager [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Took 37.35 seconds to build instance. [ 847.093214] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d90781-7149-44c3-91fc-aa6c6f859d0b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.098657] env[61867]: DEBUG nova.network.neutron [req-529a7038-8557-459a-8c0b-e418da5452cb req-2ff29609-5200-4d69-9ed5-b4a7a3ef51c2 service nova] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.115591] env[61867]: DEBUG nova.compute.provider_tree [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.160342] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276623, 'name': ReconfigVM_Task, 'duration_secs': 1.103359} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.160567] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 9aab8852-addb-49e6-a59b-fa9bffc7733b/9aab8852-addb-49e6-a59b-fa9bffc7733b.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.161363] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-71f97a23-f16b-427b-924b-b0ba709f4082 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.167798] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 847.167798] env[61867]: value = "task-1276626" [ 847.167798] env[61867]: _type = "Task" [ 847.167798] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.177471] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276626, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.364548] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276624, 'name': ReconfigVM_Task, 'duration_secs': 1.231589} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.364996] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 81e70c36-04ea-450c-9383-53ef069d1c46/81e70c36-04ea-450c-9383-53ef069d1c46.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.366031] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Updating instance '81e70c36-04ea-450c-9383-53ef069d1c46' progress to 50 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 847.387391] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276625, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.405792] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Releasing lock "refresh_cache-6e41989e-b8fa-4009-af1e-1ce859b329a1" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.590680] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2cdaf44f-f5c7-46ec-899d-479ee4275d64 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.711s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.601437] env[61867]: DEBUG oslo_concurrency.lockutils [req-529a7038-8557-459a-8c0b-e418da5452cb req-2ff29609-5200-4d69-9ed5-b4a7a3ef51c2 service nova] Releasing lock "refresh_cache-e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.601840] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquired lock "refresh_cache-e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.602212] env[61867]: DEBUG nova.network.neutron [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 847.618198] env[61867]: DEBUG nova.scheduler.client.report [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 847.679121] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276626, 'name': Rename_Task, 'duration_secs': 0.176616} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.679422] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 847.680025] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-825f85af-182d-4850-a967-36c67a4dd751 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.686824] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 847.686824] env[61867]: value = "task-1276627" [ 847.686824] env[61867]: _type = "Task" [ 847.686824] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.694497] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276627, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.715292] env[61867]: DEBUG nova.compute.manager [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 847.746761] env[61867]: DEBUG nova.virt.hardware [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.747182] env[61867]: DEBUG nova.virt.hardware [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.747591] env[61867]: DEBUG nova.virt.hardware [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.747904] env[61867]: DEBUG nova.virt.hardware [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.748129] env[61867]: DEBUG nova.virt.hardware [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.748811] env[61867]: DEBUG nova.virt.hardware [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.748998] env[61867]: DEBUG nova.virt.hardware [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.749250] env[61867]: DEBUG nova.virt.hardware [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.749511] env[61867]: DEBUG nova.virt.hardware [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.750101] env[61867]: DEBUG nova.virt.hardware [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.750976] env[61867]: DEBUG nova.virt.hardware [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.752089] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b9ca8ea-0d5e-47fe-a79a-1d90532dad53 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.763908] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f7522c7-e554-4a6d-96cf-4b90535eea42 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.872486] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee0e544-0377-4a50-9894-b9be9dc8a659 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.907749] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28895bf2-3641-4d00-bb8a-c313ed0c40ed {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.917371] env[61867]: DEBUG oslo_vmware.api [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276625, 'name': PowerOnVM_Task, 'duration_secs': 0.522298} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.933444] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 847.933613] env[61867]: INFO nova.compute.manager [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Took 10.25 seconds to spawn the instance on the hypervisor. [ 847.933799] env[61867]: DEBUG nova.compute.manager [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 847.934166] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Updating instance '81e70c36-04ea-450c-9383-53ef069d1c46' progress to 67 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 847.938162] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab0a5e85-fefe-41d0-8016-99364836bde8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.953313] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 847.957641] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4dea194a-35e8-43d2-83d3-ccc4420a176e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.964544] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 847.964544] env[61867]: value = "task-1276628" [ 847.964544] env[61867]: _type = "Task" [ 847.964544] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.978366] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276628, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.123730] env[61867]: DEBUG oslo_concurrency.lockutils [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.124304] env[61867]: DEBUG nova.compute.manager [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 848.127162] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.935s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.127399] env[61867]: DEBUG nova.objects.instance [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Lazy-loading 'resources' on Instance uuid 66518b2a-0242-438b-ba9f-d57c07a1165c {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 848.196666] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276627, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.325333] env[61867]: DEBUG nova.network.neutron [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 848.478770] env[61867]: INFO nova.compute.manager [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Took 34.57 seconds to build instance. [ 848.485983] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276628, 'name': PowerOffVM_Task, 'duration_secs': 0.244104} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.486696] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 848.487974] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd3a962-de46-465f-89f6-6db3910dcbb6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.511177] env[61867]: DEBUG nova.network.neutron [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Port 79dddbac-ec4f-4cd9-8436-a3600957e598 binding to destination host cpu-1 is already ACTIVE {{(pid=61867) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 848.514731] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16cdcadb-129f-4473-9f06-69b59169ba79 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.519054] env[61867]: DEBUG nova.compute.manager [req-580c38fd-8f75-4f32-ba22-dc14bd485680 req-450e84d1-5437-4ef6-aeb2-cffa9416486f service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Received event network-changed-af89c243-c119-47c9-9eda-60c9be2b8c51 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.519279] env[61867]: DEBUG nova.compute.manager [req-580c38fd-8f75-4f32-ba22-dc14bd485680 req-450e84d1-5437-4ef6-aeb2-cffa9416486f service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Refreshing instance network info cache due to event network-changed-af89c243-c119-47c9-9eda-60c9be2b8c51. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 848.519496] env[61867]: DEBUG oslo_concurrency.lockutils [req-580c38fd-8f75-4f32-ba22-dc14bd485680 req-450e84d1-5437-4ef6-aeb2-cffa9416486f service nova] Acquiring lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.519645] env[61867]: DEBUG oslo_concurrency.lockutils [req-580c38fd-8f75-4f32-ba22-dc14bd485680 req-450e84d1-5437-4ef6-aeb2-cffa9416486f service nova] Acquired lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.521074] env[61867]: DEBUG nova.network.neutron [req-580c38fd-8f75-4f32-ba22-dc14bd485680 req-450e84d1-5437-4ef6-aeb2-cffa9416486f service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Refreshing network info cache for port af89c243-c119-47c9-9eda-60c9be2b8c51 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 848.557193] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 848.557768] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-20fa5092-ba86-4496-9c9a-fab4c30ac118 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.566740] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 848.566740] env[61867]: value = "task-1276629" [ 848.566740] env[61867]: _type = "Task" [ 848.566740] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.575336] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] VM already powered off {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 848.575559] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 848.576301] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.576489] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.576685] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 848.576925] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f56f383c-d959-4156-8fe3-73e22f637686 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.590016] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 848.590016] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 848.590016] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90fce836-a81b-439a-8c49-752cb5bf7ead {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.594749] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 848.594749] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523a0790-7b7c-b10a-824c-effd08de7825" [ 848.594749] env[61867]: _type = "Task" [ 848.594749] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.604020] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523a0790-7b7c-b10a-824c-effd08de7825, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.607051] env[61867]: DEBUG nova.network.neutron [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Updating instance_info_cache with network_info: [{"id": "8fe6f5be-9448-4542-a731-fc42b5ab06ca", "address": "fa:16:3e:a1:41:a5", "network": {"id": "a645da41-1400-470e-acf2-68cec86432fd", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1387586368-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a29c5e9cfdaa4cc88ed300100a308ffa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fe6f5be-94", "ovs_interfaceid": "8fe6f5be-9448-4542-a731-fc42b5ab06ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.630992] env[61867]: DEBUG nova.compute.utils [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 848.636796] env[61867]: DEBUG nova.compute.manager [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 848.636991] env[61867]: DEBUG nova.network.neutron [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 848.688381] env[61867]: DEBUG nova.policy [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '86b7e5f767b94dc8a88b1e7f69c923e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '05dd2be830f9410fbb90415cc13ff6bf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 848.700914] env[61867]: DEBUG oslo_vmware.api [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276627, 'name': PowerOnVM_Task, 'duration_secs': 0.809099} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.701228] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 848.701434] env[61867]: DEBUG nova.compute.manager [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 848.702217] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6227bd1-8663-4156-b2d2-d74085c49955 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.782823] env[61867]: DEBUG nova.network.neutron [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Successfully updated port: 2b465a8c-afc2-4890-a172-d8beb1215f8a {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 848.915106] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb23160-2e82-4278-96f0-87422457fc2f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.922806] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd531a4-4894-4daa-9a16-42fa0c6b858e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.953731] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c63b55-8650-458b-a850-a4172b570c31 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.961197] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-822c7d2e-26aa-4b9b-9998-8885d1468f7d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.975318] env[61867]: DEBUG nova.compute.provider_tree [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.980166] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6de5d510-046a-46bb-ac54-656ece62f1e6 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Lock "410a2819-ea27-4613-9f2c-279f9e82a4b3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.184s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.065342] env[61867]: DEBUG nova.network.neutron [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Successfully created port: 101c4e9b-0368-4c02-9297-4eb478b6fc2b {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.106157] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523a0790-7b7c-b10a-824c-effd08de7825, 'name': SearchDatastore_Task, 'duration_secs': 0.018912} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.107323] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5369ebbe-030f-4018-8e89-620943608d10 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.110794] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Releasing lock "refresh_cache-e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.111279] env[61867]: DEBUG nova.compute.manager [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Instance network_info: |[{"id": "8fe6f5be-9448-4542-a731-fc42b5ab06ca", "address": "fa:16:3e:a1:41:a5", "network": {"id": "a645da41-1400-470e-acf2-68cec86432fd", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1387586368-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a29c5e9cfdaa4cc88ed300100a308ffa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fe6f5be-94", "ovs_interfaceid": "8fe6f5be-9448-4542-a731-fc42b5ab06ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 849.111654] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a1:41:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d377d75-3add-4a15-8691-74b2eb010924', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8fe6f5be-9448-4542-a731-fc42b5ab06ca', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 849.119802] env[61867]: DEBUG oslo.service.loopingcall [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.120829] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 849.121165] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2df462bd-5dd7-49da-b610-d30634d69736 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.140229] env[61867]: DEBUG nova.compute.manager [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 849.143474] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 849.143474] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52155868-cc48-f79c-cf9d-bf35748029d5" [ 849.143474] env[61867]: _type = "Task" [ 849.143474] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.152049] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.152049] env[61867]: value = "task-1276630" [ 849.152049] env[61867]: _type = "Task" [ 849.152049] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.156305] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52155868-cc48-f79c-cf9d-bf35748029d5, 'name': SearchDatastore_Task, 'duration_secs': 0.025279} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.159725] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.159725] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 6e41989e-b8fa-4009-af1e-1ce859b329a1/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk. {{(pid=61867) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 849.160042] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2bdb3452-2191-44cc-ae2c-9398d6330508 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.166982] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276630, 'name': CreateVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.168382] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 849.168382] env[61867]: value = "task-1276631" [ 849.168382] env[61867]: _type = "Task" [ 849.168382] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.175892] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276631, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.223315] env[61867]: DEBUG oslo_concurrency.lockutils [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.286512] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "refresh_cache-71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.286671] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "refresh_cache-71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.286874] env[61867]: DEBUG nova.network.neutron [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 849.415692] env[61867]: DEBUG nova.network.neutron [req-580c38fd-8f75-4f32-ba22-dc14bd485680 req-450e84d1-5437-4ef6-aeb2-cffa9416486f service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Updated VIF entry in instance network info cache for port af89c243-c119-47c9-9eda-60c9be2b8c51. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 849.415959] env[61867]: DEBUG nova.network.neutron [req-580c38fd-8f75-4f32-ba22-dc14bd485680 req-450e84d1-5437-4ef6-aeb2-cffa9416486f service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Updating instance_info_cache with network_info: [{"id": "af89c243-c119-47c9-9eda-60c9be2b8c51", "address": "fa:16:3e:4d:1f:76", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf89c243-c1", "ovs_interfaceid": "af89c243-c119-47c9-9eda-60c9be2b8c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.482031] env[61867]: DEBUG nova.scheduler.client.report [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 849.546605] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "81e70c36-04ea-450c-9383-53ef069d1c46-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.546605] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "81e70c36-04ea-450c-9383-53ef069d1c46-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.546749] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "81e70c36-04ea-450c-9383-53ef069d1c46-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.671830] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276630, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.682169] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276631, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.820921] env[61867]: DEBUG nova.network.neutron [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 849.919849] env[61867]: DEBUG oslo_concurrency.lockutils [req-580c38fd-8f75-4f32-ba22-dc14bd485680 req-450e84d1-5437-4ef6-aeb2-cffa9416486f service nova] Releasing lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.920371] env[61867]: INFO nova.compute.manager [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Rescuing [ 849.920644] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Acquiring lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.920985] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Acquired lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.921229] env[61867]: DEBUG nova.network.neutron [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 849.958571] env[61867]: DEBUG nova.network.neutron [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Updating instance_info_cache with network_info: [{"id": "2b465a8c-afc2-4890-a172-d8beb1215f8a", "address": "fa:16:3e:d6:d4:86", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b465a8c-af", "ovs_interfaceid": "2b465a8c-afc2-4890-a172-d8beb1215f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.987516] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "9aab8852-addb-49e6-a59b-fa9bffc7733b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.987776] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "9aab8852-addb-49e6-a59b-fa9bffc7733b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.988007] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "9aab8852-addb-49e6-a59b-fa9bffc7733b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.988219] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "9aab8852-addb-49e6-a59b-fa9bffc7733b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.988394] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "9aab8852-addb-49e6-a59b-fa9bffc7733b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.990539] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.863s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.992824] env[61867]: INFO nova.compute.manager [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Terminating instance [ 849.995525] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.613s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.996577] env[61867]: INFO nova.compute.claims [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 849.999435] env[61867]: DEBUG nova.compute.manager [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 849.999664] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 850.000525] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a3933f7-2310-4312-92e7-296d7a619d5b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.008791] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 850.009069] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e5b6bbc-e4bd-467f-b766-de40d3db9367 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.015930] env[61867]: DEBUG oslo_vmware.api [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 850.015930] env[61867]: value = "task-1276632" [ 850.015930] env[61867]: _type = "Task" [ 850.015930] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.019886] env[61867]: INFO nova.scheduler.client.report [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Deleted allocations for instance 66518b2a-0242-438b-ba9f-d57c07a1165c [ 850.027384] env[61867]: DEBUG oslo_vmware.api [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276632, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.154053] env[61867]: DEBUG nova.compute.manager [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 850.169558] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276630, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.180456] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276631, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595559} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.182682] env[61867]: INFO nova.virt.vmwareapi.ds_util [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 6e41989e-b8fa-4009-af1e-1ce859b329a1/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk. [ 850.183655] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c18be1bb-8467-4a7a-bac3-126fc65c05fe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.190116] env[61867]: DEBUG nova.virt.hardware [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 850.190116] env[61867]: DEBUG nova.virt.hardware [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 850.190116] env[61867]: DEBUG nova.virt.hardware [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.190279] env[61867]: DEBUG nova.virt.hardware [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 850.190279] env[61867]: DEBUG nova.virt.hardware [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.190279] env[61867]: DEBUG nova.virt.hardware [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 850.190402] env[61867]: DEBUG nova.virt.hardware [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 850.190565] env[61867]: DEBUG nova.virt.hardware [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 850.190765] env[61867]: DEBUG nova.virt.hardware [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 850.190940] env[61867]: DEBUG nova.virt.hardware [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 850.191134] env[61867]: DEBUG nova.virt.hardware [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 850.192103] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63bf4444-8152-458a-a9fe-bc18c3861b8d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.216207] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 6e41989e-b8fa-4009-af1e-1ce859b329a1/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 850.217057] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c93fdc40-685c-44f4-8aa7-88444201cf74 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.232104] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "1df8427c-e75d-4b60-a92a-b5ba76b67081" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.232356] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "1df8427c-e75d-4b60-a92a-b5ba76b67081" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.232535] env[61867]: INFO nova.compute.manager [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Shelving [ 850.236363] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72866bd5-e42e-4920-9698-7b75fdf338b7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.244292] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 850.244292] env[61867]: value = "task-1276633" [ 850.244292] env[61867]: _type = "Task" [ 850.244292] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.263854] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276633, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.461124] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "refresh_cache-71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.461471] env[61867]: DEBUG nova.compute.manager [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Instance network_info: |[{"id": "2b465a8c-afc2-4890-a172-d8beb1215f8a", "address": "fa:16:3e:d6:d4:86", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b465a8c-af", "ovs_interfaceid": "2b465a8c-afc2-4890-a172-d8beb1215f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 850.461927] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d6:d4:86', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74e6f6e0-95e6-4531-99e9-0e78350fb655', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2b465a8c-afc2-4890-a172-d8beb1215f8a', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.470301] env[61867]: DEBUG oslo.service.loopingcall [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.470584] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 850.470925] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2603b2e7-b99f-4f29-b622-803c1fde32d9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.490822] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.490822] env[61867]: value = "task-1276634" [ 850.490822] env[61867]: _type = "Task" [ 850.490822] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.501861] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276634, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.531446] env[61867]: DEBUG oslo_vmware.api [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276632, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.531921] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a01437bd-4020-4f6e-ba5b-ccd0e1a16d80 tempest-ServerMetadataTestJSON-73131360 tempest-ServerMetadataTestJSON-73131360-project-member] Lock "66518b2a-0242-438b-ba9f-d57c07a1165c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.568s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.557709] env[61867]: DEBUG nova.compute.manager [req-5be91efe-75c7-43de-b6ee-5d7385f316f0 req-116bc0f0-cc4c-4352-b584-89f9fc978002 service nova] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Received event network-vif-plugged-2b465a8c-afc2-4890-a172-d8beb1215f8a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.558037] env[61867]: DEBUG oslo_concurrency.lockutils [req-5be91efe-75c7-43de-b6ee-5d7385f316f0 req-116bc0f0-cc4c-4352-b584-89f9fc978002 service nova] Acquiring lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.558264] env[61867]: DEBUG oslo_concurrency.lockutils [req-5be91efe-75c7-43de-b6ee-5d7385f316f0 req-116bc0f0-cc4c-4352-b584-89f9fc978002 service nova] Lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.558443] env[61867]: DEBUG oslo_concurrency.lockutils [req-5be91efe-75c7-43de-b6ee-5d7385f316f0 req-116bc0f0-cc4c-4352-b584-89f9fc978002 service nova] Lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.559029] env[61867]: DEBUG nova.compute.manager [req-5be91efe-75c7-43de-b6ee-5d7385f316f0 req-116bc0f0-cc4c-4352-b584-89f9fc978002 service nova] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] No waiting events found dispatching network-vif-plugged-2b465a8c-afc2-4890-a172-d8beb1215f8a {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 850.559187] env[61867]: WARNING nova.compute.manager [req-5be91efe-75c7-43de-b6ee-5d7385f316f0 req-116bc0f0-cc4c-4352-b584-89f9fc978002 service nova] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Received unexpected event network-vif-plugged-2b465a8c-afc2-4890-a172-d8beb1215f8a for instance with vm_state building and task_state spawning. [ 850.559355] env[61867]: DEBUG nova.compute.manager [req-5be91efe-75c7-43de-b6ee-5d7385f316f0 req-116bc0f0-cc4c-4352-b584-89f9fc978002 service nova] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Received event network-changed-2b465a8c-afc2-4890-a172-d8beb1215f8a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.560079] env[61867]: DEBUG nova.compute.manager [req-5be91efe-75c7-43de-b6ee-5d7385f316f0 req-116bc0f0-cc4c-4352-b584-89f9fc978002 service nova] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Refreshing instance network info cache due to event network-changed-2b465a8c-afc2-4890-a172-d8beb1215f8a. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 850.560240] env[61867]: DEBUG oslo_concurrency.lockutils [req-5be91efe-75c7-43de-b6ee-5d7385f316f0 req-116bc0f0-cc4c-4352-b584-89f9fc978002 service nova] Acquiring lock "refresh_cache-71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.560431] env[61867]: DEBUG oslo_concurrency.lockutils [req-5be91efe-75c7-43de-b6ee-5d7385f316f0 req-116bc0f0-cc4c-4352-b584-89f9fc978002 service nova] Acquired lock "refresh_cache-71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.560723] env[61867]: DEBUG nova.network.neutron [req-5be91efe-75c7-43de-b6ee-5d7385f316f0 req-116bc0f0-cc4c-4352-b584-89f9fc978002 service nova] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Refreshing network info cache for port 2b465a8c-afc2-4890-a172-d8beb1215f8a {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 850.609949] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.610189] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquired lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.610403] env[61867]: DEBUG nova.network.neutron [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 850.667670] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276630, 'name': CreateVM_Task, 'duration_secs': 1.418376} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.667873] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 850.668679] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/420f5356-3ba1-4114-af3a-5be2a65b82cb" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.668831] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquired lock "[datastore2] devstack-image-cache_base/420f5356-3ba1-4114-af3a-5be2a65b82cb" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.669258] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/420f5356-3ba1-4114-af3a-5be2a65b82cb" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 850.669543] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77c1ce65-6fe8-4d6d-8e0f-96aff6e5e578 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.674596] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 850.674596] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e1b935-acab-af33-7171-75da541ec12f" [ 850.674596] env[61867]: _type = "Task" [ 850.674596] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.684562] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e1b935-acab-af33-7171-75da541ec12f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.719371] env[61867]: DEBUG nova.network.neutron [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Updating instance_info_cache with network_info: [{"id": "cf5eac56-8386-49e1-a2b3-d19e3d8297d6", "address": "fa:16:3e:a9:7a:17", "network": {"id": "06608a13-6187-41bb-926a-1fbda4f627a0", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-618264839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e54d4ce9d6c24e8ab40cb77886aac634", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf5eac56-83", "ovs_interfaceid": "cf5eac56-8386-49e1-a2b3-d19e3d8297d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.740257] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 850.740523] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96375fe1-a0be-4596-8f3e-57d898a7532d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.751785] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 850.751785] env[61867]: value = "task-1276635" [ 850.751785] env[61867]: _type = "Task" [ 850.751785] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.759943] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276633, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.764991] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276635, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.778996] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "07fdd75c-34ab-45e5-a98b-aced2caa6cb5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.779307] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "07fdd75c-34ab-45e5-a98b-aced2caa6cb5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.876363] env[61867]: DEBUG nova.network.neutron [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Successfully updated port: 101c4e9b-0368-4c02-9297-4eb478b6fc2b {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 851.001201] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276634, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.029819] env[61867]: DEBUG oslo_vmware.api [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276632, 'name': PowerOffVM_Task, 'duration_secs': 0.751844} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.032873] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 851.033093] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 851.033784] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9323c94-b4ab-455d-a449-976827d8889e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.188953] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Releasing lock "[datastore2] devstack-image-cache_base/420f5356-3ba1-4114-af3a-5be2a65b82cb" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.189849] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Processing image 420f5356-3ba1-4114-af3a-5be2a65b82cb {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 851.189849] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/420f5356-3ba1-4114-af3a-5be2a65b82cb/420f5356-3ba1-4114-af3a-5be2a65b82cb.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.189849] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquired lock "[datastore2] devstack-image-cache_base/420f5356-3ba1-4114-af3a-5be2a65b82cb/420f5356-3ba1-4114-af3a-5be2a65b82cb.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.190126] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 851.192334] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4837aea9-f53e-429b-8bb6-ba2400184846 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.208069] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 851.208069] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 851.208069] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a59f609-25bc-46c4-8728-2f1cbe578862 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.212913] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 851.212913] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528afa7e-1d48-1248-698f-ea8c1e006047" [ 851.212913] env[61867]: _type = "Task" [ 851.212913] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.225019] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Releasing lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.227266] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528afa7e-1d48-1248-698f-ea8c1e006047, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.261072] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276633, 'name': ReconfigVM_Task, 'duration_secs': 0.642445} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.270283] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 6e41989e-b8fa-4009-af1e-1ce859b329a1/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 851.277073] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3493bfd-cbd5-4747-b751-d5501fdfffc1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.279575] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276635, 'name': PowerOffVM_Task, 'duration_secs': 0.349309} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.282821] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 851.283246] env[61867]: DEBUG nova.compute.manager [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 851.286920] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf3fa455-bb16-4c46-ae68-1db6dab5f52d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.316465] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5bd12923-665b-4b25-9ec9-324117e68460 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.346162] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9158157-a419-4faf-af46-cb2d24f17bfa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.351248] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb80e942-52ff-4aec-9ea6-639aed3322aa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.353676] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 851.353676] env[61867]: value = "task-1276637" [ 851.353676] env[61867]: _type = "Task" [ 851.353676] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.363507] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67941e06-1b13-454a-9301-a35725fd2579 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.370934] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276637, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.397678] env[61867]: DEBUG oslo_concurrency.lockutils [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "refresh_cache-6d2dab88-4165-4952-8019-2eaf3b863115" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.397832] env[61867]: DEBUG oslo_concurrency.lockutils [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquired lock "refresh_cache-6d2dab88-4165-4952-8019-2eaf3b863115" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.398109] env[61867]: DEBUG nova.network.neutron [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 851.403192] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd4f871a-91b0-4e87-a882-f4bc4dff64b0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.406994] env[61867]: DEBUG nova.network.neutron [req-5be91efe-75c7-43de-b6ee-5d7385f316f0 req-116bc0f0-cc4c-4352-b584-89f9fc978002 service nova] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Updated VIF entry in instance network info cache for port 2b465a8c-afc2-4890-a172-d8beb1215f8a. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 851.407353] env[61867]: DEBUG nova.network.neutron [req-5be91efe-75c7-43de-b6ee-5d7385f316f0 req-116bc0f0-cc4c-4352-b584-89f9fc978002 service nova] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Updating instance_info_cache with network_info: [{"id": "2b465a8c-afc2-4890-a172-d8beb1215f8a", "address": "fa:16:3e:d6:d4:86", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b465a8c-af", "ovs_interfaceid": "2b465a8c-afc2-4890-a172-d8beb1215f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.415283] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3af892f-5ad6-4c87-ad55-4a81f7f3e4d3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.419623] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 851.419822] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 851.420060] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleting the datastore file [datastore1] 9aab8852-addb-49e6-a59b-fa9bffc7733b {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 851.420931] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e06f2427-f42d-472d-83c5-b6e98d059d03 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.432573] env[61867]: DEBUG nova.compute.provider_tree [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.434736] env[61867]: DEBUG oslo_vmware.api [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 851.434736] env[61867]: value = "task-1276638" [ 851.434736] env[61867]: _type = "Task" [ 851.434736] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.442781] env[61867]: DEBUG oslo_vmware.api [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276638, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.479298] env[61867]: DEBUG nova.network.neutron [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Updating instance_info_cache with network_info: [{"id": "79dddbac-ec4f-4cd9-8436-a3600957e598", "address": "fa:16:3e:b8:10:25", "network": {"id": "2d4ab52c-144f-45a1-9d4f-afd85fc30404", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.63", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "84601e35c8e8487cb78fc16a2536a4c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d47d5e1d-e66d-4f2c-83e6-d5e78c2b767d", "external-id": "nsx-vlan-transportzone-109", "segmentation_id": 109, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79dddbac-ec", "ovs_interfaceid": "79dddbac-ec4f-4cd9-8436-a3600957e598", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.501281] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276634, 'name': CreateVM_Task, 'duration_secs': 0.631158} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.501452] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 851.502114] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.502292] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.502609] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.502864] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0dd0990-b386-4e9a-8e2a-54a112dee163 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.507571] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 851.507571] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5200c3af-e079-3f09-5d37-7eaeac79486c" [ 851.507571] env[61867]: _type = "Task" [ 851.507571] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.515392] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5200c3af-e079-3f09-5d37-7eaeac79486c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.725022] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Preparing fetch location {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 851.725022] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Fetch image to [datastore2] OSTACK_IMG_fc57df4d-cf10-49c0-a1d1-738dd6e7ffda/OSTACK_IMG_fc57df4d-cf10-49c0-a1d1-738dd6e7ffda.vmdk {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 851.725022] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Downloading stream optimized image 420f5356-3ba1-4114-af3a-5be2a65b82cb to [datastore2] OSTACK_IMG_fc57df4d-cf10-49c0-a1d1-738dd6e7ffda/OSTACK_IMG_fc57df4d-cf10-49c0-a1d1-738dd6e7ffda.vmdk on the data store datastore2 as vApp {{(pid=61867) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 851.725022] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Downloading image file data 420f5356-3ba1-4114-af3a-5be2a65b82cb to the ESX as VM named 'OSTACK_IMG_fc57df4d-cf10-49c0-a1d1-738dd6e7ffda' {{(pid=61867) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 851.781068] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 851.781788] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-edaa4531-b0bb-4463-bc9b-ef2224ef7938 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.789536] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 851.789536] env[61867]: value = "task-1276639" [ 851.789536] env[61867]: _type = "Task" [ 851.789536] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.802460] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276639, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.805819] env[61867]: DEBUG oslo_vmware.rw_handles [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 851.805819] env[61867]: value = "resgroup-9" [ 851.805819] env[61867]: _type = "ResourcePool" [ 851.805819] env[61867]: }. {{(pid=61867) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 851.806352] env[61867]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-d213c6b9-f4b5-48f9-aa55-6c8c15a7a88b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.823032] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.830043] env[61867]: DEBUG oslo_vmware.rw_handles [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lease: (returnval){ [ 851.830043] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5210cee9-f2c4-700d-e8dd-82d663201ded" [ 851.830043] env[61867]: _type = "HttpNfcLease" [ 851.830043] env[61867]: } obtained for vApp import into resource pool (val){ [ 851.830043] env[61867]: value = "resgroup-9" [ 851.830043] env[61867]: _type = "ResourcePool" [ 851.830043] env[61867]: }. {{(pid=61867) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 851.830719] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the lease: (returnval){ [ 851.830719] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5210cee9-f2c4-700d-e8dd-82d663201ded" [ 851.830719] env[61867]: _type = "HttpNfcLease" [ 851.830719] env[61867]: } to be ready. {{(pid=61867) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 851.838516] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 851.838516] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5210cee9-f2c4-700d-e8dd-82d663201ded" [ 851.838516] env[61867]: _type = "HttpNfcLease" [ 851.838516] env[61867]: } is initializing. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 851.863622] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276637, 'name': ReconfigVM_Task, 'duration_secs': 0.51275} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.864087] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 851.864453] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a3b9616-a689-418c-9324-b1e8e17d7475 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.871253] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Creating Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 851.871253] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ee7d036c-f629-47ef-8983-88ba3e7b74e3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.873863] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 851.873863] env[61867]: value = "task-1276641" [ 851.873863] env[61867]: _type = "Task" [ 851.873863] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.878519] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 851.878519] env[61867]: value = "task-1276642" [ 851.878519] env[61867]: _type = "Task" [ 851.878519] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.885918] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276641, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.892176] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276642, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.909968] env[61867]: DEBUG oslo_concurrency.lockutils [req-5be91efe-75c7-43de-b6ee-5d7385f316f0 req-116bc0f0-cc4c-4352-b584-89f9fc978002 service nova] Releasing lock "refresh_cache-71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.936323] env[61867]: DEBUG nova.scheduler.client.report [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.949805] env[61867]: DEBUG oslo_vmware.api [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276638, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.466039} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.950050] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 851.950128] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 851.950278] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 851.950525] env[61867]: INFO nova.compute.manager [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Took 1.95 seconds to destroy the instance on the hypervisor. [ 851.950727] env[61867]: DEBUG oslo.service.loopingcall [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.950943] env[61867]: DEBUG nova.compute.manager [-] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 851.951044] env[61867]: DEBUG nova.network.neutron [-] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 851.980126] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Releasing lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.999903] env[61867]: DEBUG nova.network.neutron [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.019580] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5200c3af-e079-3f09-5d37-7eaeac79486c, 'name': SearchDatastore_Task, 'duration_secs': 0.020198} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.024105] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.024105] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.024105] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.024105] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.024229] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.024229] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b75c09b-174a-4263-b69f-0aa91d9e747b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.041865] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.042312] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 852.043854] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a5d6373-b47b-4a90-9432-8625c94e9e59 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.050581] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 852.050581] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b0d97a-00de-1682-9f3d-d458024ae77f" [ 852.050581] env[61867]: _type = "Task" [ 852.050581] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.060033] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b0d97a-00de-1682-9f3d-d458024ae77f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.311441] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276639, 'name': PowerOffVM_Task, 'duration_secs': 0.210399} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.313248] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 852.313248] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762c9e65-48a8-4e87-9cc7-9cd7516c1243 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.338904] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-835adb02-7682-48a0-a3cf-c9d56645e496 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.354622] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 852.354622] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5210cee9-f2c4-700d-e8dd-82d663201ded" [ 852.354622] env[61867]: _type = "HttpNfcLease" [ 852.354622] env[61867]: } is ready. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 852.355862] env[61867]: DEBUG nova.network.neutron [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Updating instance_info_cache with network_info: [{"id": "101c4e9b-0368-4c02-9297-4eb478b6fc2b", "address": "fa:16:3e:50:da:3d", "network": {"id": "9c6e0d78-8bbe-46b4-9233-94f991e940e8", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1003528124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05dd2be830f9410fbb90415cc13ff6bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap101c4e9b-03", "ovs_interfaceid": "101c4e9b-0368-4c02-9297-4eb478b6fc2b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.358244] env[61867]: DEBUG oslo_vmware.rw_handles [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 852.358244] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5210cee9-f2c4-700d-e8dd-82d663201ded" [ 852.358244] env[61867]: _type = "HttpNfcLease" [ 852.358244] env[61867]: }. {{(pid=61867) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 852.359184] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-264a85fa-e9dd-4233-bbf7-e803607bb146 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.367731] env[61867]: DEBUG oslo_vmware.rw_handles [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52728fb1-cfad-e7a4-4cac-a9985019b44b/disk-0.vmdk from lease info. {{(pid=61867) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 852.367974] env[61867]: DEBUG oslo_vmware.rw_handles [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52728fb1-cfad-e7a4-4cac-a9985019b44b/disk-0.vmdk. {{(pid=61867) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 852.448539] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-860b191e-6688-44c7-8976-9d391db53489 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.450575] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.451093] env[61867]: DEBUG nova.compute.manager [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 852.454274] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276642, 'name': CreateSnapshot_Task, 'duration_secs': 0.490409} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.454848] env[61867]: DEBUG oslo_vmware.api [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276641, 'name': PowerOnVM_Task, 'duration_secs': 0.408658} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.457254] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 852.459678] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.340s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.460616] env[61867]: INFO nova.compute.claims [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 852.463033] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Created Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 852.463326] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 852.464833] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-027d1c7b-c618-409a-994c-e27f260c0273 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.467400] env[61867]: DEBUG nova.compute.manager [None req-d88a3d96-c757-4dd0-91c7-6853e28756bd tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.468901] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2669273-2cc8-4a8d-855f-e9fc37d63a3e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.473926] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fd01bcb-8c27-439f-9400-4528bf15ae37 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.492060] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 852.492060] env[61867]: value = "task-1276643" [ 852.492060] env[61867]: _type = "Task" [ 852.492060] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.503055] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] VM already powered off {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 852.503055] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.503978] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.503978] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.503978] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.503978] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3c3971e3-4750-43a3-b176-905f0b371449 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.513259] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.513431] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 852.514443] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf85b0e0-c7e8-46f7-ac75-cbe92690289a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.518762] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a102c58c-4e09-4df6-a66a-18caab952371 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.524953] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 852.524953] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526d115d-d911-dd6b-4031-e883e695bd87" [ 852.524953] env[61867]: _type = "Task" [ 852.524953] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.543208] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63570b69-4b55-4feb-a585-ae1d10592250 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.556910] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526d115d-d911-dd6b-4031-e883e695bd87, 'name': SearchDatastore_Task, 'duration_secs': 0.010845} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.557281] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Updating instance '81e70c36-04ea-450c-9383-53ef069d1c46' progress to 83 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 852.566029] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79260130-58e5-4851-9d97-41ca81f2ac8e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.572680] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b0d97a-00de-1682-9f3d-d458024ae77f, 'name': SearchDatastore_Task, 'duration_secs': 0.008925} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.574318] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 852.574318] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521807e4-217b-4b67-e48f-21c291f7ab2e" [ 852.574318] env[61867]: _type = "Task" [ 852.574318] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.574552] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1124273b-dfa4-4c32-8930-24ee912a2661 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.582309] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 852.582309] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5273eca8-9840-2a24-6e39-a019358faaa6" [ 852.582309] env[61867]: _type = "Task" [ 852.582309] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.587055] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521807e4-217b-4b67-e48f-21c291f7ab2e, 'name': SearchDatastore_Task, 'duration_secs': 0.009804} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.588837] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.589355] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 410a2819-ea27-4613-9f2c-279f9e82a4b3/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk. {{(pid=61867) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 852.589456] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4d847fb9-adce-470b-9364-e8490472f9c3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.598346] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5273eca8-9840-2a24-6e39-a019358faaa6, 'name': SearchDatastore_Task, 'duration_secs': 0.01109} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.598346] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.598346] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe/71cd9036-0f99-4e30-aad6-ceb4f15d4ffe.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 852.598346] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 852.598346] env[61867]: value = "task-1276644" [ 852.598346] env[61867]: _type = "Task" [ 852.598346] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.598909] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5efa88c2-85f0-4708-9c5b-1521ee63620e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.608583] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276644, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.609870] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 852.609870] env[61867]: value = "task-1276645" [ 852.609870] env[61867]: _type = "Task" [ 852.609870] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.617959] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276645, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.647417] env[61867]: DEBUG nova.compute.manager [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Received event network-vif-plugged-101c4e9b-0368-4c02-9297-4eb478b6fc2b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.647558] env[61867]: DEBUG oslo_concurrency.lockutils [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] Acquiring lock "6d2dab88-4165-4952-8019-2eaf3b863115-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.647778] env[61867]: DEBUG oslo_concurrency.lockutils [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] Lock "6d2dab88-4165-4952-8019-2eaf3b863115-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.647979] env[61867]: DEBUG oslo_concurrency.lockutils [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] Lock "6d2dab88-4165-4952-8019-2eaf3b863115-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.648181] env[61867]: DEBUG nova.compute.manager [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] No waiting events found dispatching network-vif-plugged-101c4e9b-0368-4c02-9297-4eb478b6fc2b {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 852.648361] env[61867]: WARNING nova.compute.manager [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Received unexpected event network-vif-plugged-101c4e9b-0368-4c02-9297-4eb478b6fc2b for instance with vm_state building and task_state spawning. [ 852.648542] env[61867]: DEBUG nova.compute.manager [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Received event network-changed-101c4e9b-0368-4c02-9297-4eb478b6fc2b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.648735] env[61867]: DEBUG nova.compute.manager [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Refreshing instance network info cache due to event network-changed-101c4e9b-0368-4c02-9297-4eb478b6fc2b. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 852.652024] env[61867]: DEBUG oslo_concurrency.lockutils [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] Acquiring lock "refresh_cache-6d2dab88-4165-4952-8019-2eaf3b863115" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.863027] env[61867]: DEBUG oslo_concurrency.lockutils [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Releasing lock "refresh_cache-6d2dab88-4165-4952-8019-2eaf3b863115" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.863027] env[61867]: DEBUG nova.compute.manager [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Instance network_info: |[{"id": "101c4e9b-0368-4c02-9297-4eb478b6fc2b", "address": "fa:16:3e:50:da:3d", "network": {"id": "9c6e0d78-8bbe-46b4-9233-94f991e940e8", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1003528124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05dd2be830f9410fbb90415cc13ff6bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap101c4e9b-03", "ovs_interfaceid": "101c4e9b-0368-4c02-9297-4eb478b6fc2b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 852.863578] env[61867]: DEBUG oslo_concurrency.lockutils [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] Acquired lock "refresh_cache-6d2dab88-4165-4952-8019-2eaf3b863115" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.863578] env[61867]: DEBUG nova.network.neutron [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Refreshing network info cache for port 101c4e9b-0368-4c02-9297-4eb478b6fc2b {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 852.865123] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:50:da:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '101c4e9b-0368-4c02-9297-4eb478b6fc2b', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 852.877993] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Creating folder: Project (05dd2be830f9410fbb90415cc13ff6bf). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 852.893232] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b86c3078-7617-4ad2-9d82-62a708b47fe4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.912024] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Created folder: Project (05dd2be830f9410fbb90415cc13ff6bf) in parent group-v274258. [ 852.912024] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Creating folder: Instances. Parent ref: group-v274353. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 852.912024] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-78eea505-5ae5-4fd3-b928-ec1875f3d199 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.922295] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Created folder: Instances in parent group-v274353. [ 852.922873] env[61867]: DEBUG oslo.service.loopingcall [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.923372] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 852.923867] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c0175320-3c8e-4438-bb18-fa6e822fb89c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.957080] env[61867]: DEBUG nova.compute.utils [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 852.961200] env[61867]: DEBUG nova.compute.manager [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 852.961717] env[61867]: DEBUG nova.network.neutron [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 852.964043] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 852.964043] env[61867]: value = "task-1276648" [ 852.964043] env[61867]: _type = "Task" [ 852.964043] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.983173] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276648, 'name': CreateVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.000661] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Creating linked-clone VM from snapshot {{(pid=61867) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 853.011033] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-00a2d028-1514-487f-9ed5-a61113a76af6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.016707] env[61867]: DEBUG nova.network.neutron [-] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.027818] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 853.027818] env[61867]: value = "task-1276649" [ 853.027818] env[61867]: _type = "Task" [ 853.027818] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.047244] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276649, 'name': CloneVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.067860] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 853.070603] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d2aa8f4-3e60-4b46-897b-4578b234ce14 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.079330] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 853.079330] env[61867]: value = "task-1276650" [ 853.079330] env[61867]: _type = "Task" [ 853.079330] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.095531] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276650, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.112856] env[61867]: DEBUG nova.policy [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '322fe9fc12a34183b4241e28f7254df4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24180ce8aca142fb897e29d4853c20c3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 853.124759] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276644, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.133027] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276645, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.374496] env[61867]: DEBUG oslo_vmware.rw_handles [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Completed reading data from the image iterator. {{(pid=61867) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 853.374600] env[61867]: DEBUG oslo_vmware.rw_handles [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52728fb1-cfad-e7a4-4cac-a9985019b44b/disk-0.vmdk. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 853.377499] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87075f5d-7856-4744-aaa0-16d13f5bc552 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.384321] env[61867]: DEBUG oslo_vmware.rw_handles [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52728fb1-cfad-e7a4-4cac-a9985019b44b/disk-0.vmdk is in state: ready. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 853.384516] env[61867]: DEBUG oslo_vmware.rw_handles [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52728fb1-cfad-e7a4-4cac-a9985019b44b/disk-0.vmdk. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 853.384798] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-cd57bceb-dda0-4439-9449-6c4efc1c3bf9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.413406] env[61867]: DEBUG nova.network.neutron [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Updated VIF entry in instance network info cache for port 101c4e9b-0368-4c02-9297-4eb478b6fc2b. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 853.413406] env[61867]: DEBUG nova.network.neutron [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Updating instance_info_cache with network_info: [{"id": "101c4e9b-0368-4c02-9297-4eb478b6fc2b", "address": "fa:16:3e:50:da:3d", "network": {"id": "9c6e0d78-8bbe-46b4-9233-94f991e940e8", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1003528124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05dd2be830f9410fbb90415cc13ff6bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap101c4e9b-03", "ovs_interfaceid": "101c4e9b-0368-4c02-9297-4eb478b6fc2b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.465998] env[61867]: DEBUG nova.compute.manager [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 853.482579] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276648, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.521637] env[61867]: INFO nova.compute.manager [-] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Took 1.57 seconds to deallocate network for instance. [ 853.546630] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276649, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.597803] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276650, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.617928] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276644, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.636985} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.622099] env[61867]: INFO nova.virt.vmwareapi.ds_util [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 410a2819-ea27-4613-9f2c-279f9e82a4b3/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk. [ 853.622334] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664d939b-0787-47ec-a83c-c9bbacf8fc94 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.632933] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276645, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.63736} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.648592] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe/71cd9036-0f99-4e30-aad6-ceb4f15d4ffe.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 853.648914] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 853.657242] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 410a2819-ea27-4613-9f2c-279f9e82a4b3/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 853.660137] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bea91e88-80a5-40b8-937c-a51e16daa063 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.662235] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-774e1e2d-dbee-4518-8f2e-005f709e0fba {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.683161] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 853.683161] env[61867]: value = "task-1276652" [ 853.683161] env[61867]: _type = "Task" [ 853.683161] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.684671] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 853.684671] env[61867]: value = "task-1276651" [ 853.684671] env[61867]: _type = "Task" [ 853.684671] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.699290] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276652, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.702374] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276651, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.707744] env[61867]: DEBUG nova.network.neutron [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Successfully created port: a3a61d3e-16b8-476c-8ff4-fb229c392f72 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 853.747635] env[61867]: DEBUG oslo_vmware.rw_handles [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52728fb1-cfad-e7a4-4cac-a9985019b44b/disk-0.vmdk. {{(pid=61867) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 853.748271] env[61867]: INFO nova.virt.vmwareapi.images [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Downloaded image file data 420f5356-3ba1-4114-af3a-5be2a65b82cb [ 853.749163] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77fc7a7a-6e53-4bbc-bc6b-9b95ed37c397 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.767846] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4cc0b547-6857-48dc-973e-4a001a7474f7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.821419] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56f6bac-6102-493b-9d07-d9535f1c22e0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.827578] env[61867]: INFO nova.virt.vmwareapi.images [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] The imported VM was unregistered [ 853.828164] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Caching image {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 853.828401] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Creating directory with path [datastore2] devstack-image-cache_base/420f5356-3ba1-4114-af3a-5be2a65b82cb {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 853.829121] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-33ffc28f-9621-412c-bddd-0e67453ceb88 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.834663] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-038ca069-0729-4efc-9a3f-34e229cd1654 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.865855] env[61867]: INFO nova.compute.manager [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Unrescuing [ 853.866143] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "refresh_cache-6e41989e-b8fa-4009-af1e-1ce859b329a1" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.866281] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired lock "refresh_cache-6e41989e-b8fa-4009-af1e-1ce859b329a1" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.866442] env[61867]: DEBUG nova.network.neutron [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 853.869500] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d35b768-0ef5-4673-952a-8d846df81603 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.873057] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Created directory with path [datastore2] devstack-image-cache_base/420f5356-3ba1-4114-af3a-5be2a65b82cb {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 853.873057] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_fc57df4d-cf10-49c0-a1d1-738dd6e7ffda/OSTACK_IMG_fc57df4d-cf10-49c0-a1d1-738dd6e7ffda.vmdk to [datastore2] devstack-image-cache_base/420f5356-3ba1-4114-af3a-5be2a65b82cb/420f5356-3ba1-4114-af3a-5be2a65b82cb.vmdk. {{(pid=61867) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 853.873237] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-d91b188c-6a7d-4a35-852d-2580c35a502d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.882502] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a0915b-16cf-49a8-93df-63b56aa9a5a0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.888260] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 853.888260] env[61867]: value = "task-1276654" [ 853.888260] env[61867]: _type = "Task" [ 853.888260] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.901037] env[61867]: DEBUG nova.compute.provider_tree [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.907218] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276654, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.916214] env[61867]: DEBUG oslo_concurrency.lockutils [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] Releasing lock "refresh_cache-6d2dab88-4165-4952-8019-2eaf3b863115" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.916479] env[61867]: DEBUG nova.compute.manager [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Received event network-vif-deleted-104495f6-b976-4f53-9959-d5193f833ae8 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.916660] env[61867]: INFO nova.compute.manager [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Neutron deleted interface 104495f6-b976-4f53-9959-d5193f833ae8; detaching it from the instance and deleting it from the info cache [ 853.916907] env[61867]: DEBUG nova.network.neutron [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.981900] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276648, 'name': CreateVM_Task, 'duration_secs': 0.72702} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.982331] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 853.983133] env[61867]: DEBUG oslo_concurrency.lockutils [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.983315] env[61867]: DEBUG oslo_concurrency.lockutils [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.983658] env[61867]: DEBUG oslo_concurrency.lockutils [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 853.983935] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-682cf0d9-9c37-4539-8594-ef0ef388b41a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.988923] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 853.988923] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5298158b-22df-d1f3-7e39-8401141a38d6" [ 853.988923] env[61867]: _type = "Task" [ 853.988923] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.997484] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5298158b-22df-d1f3-7e39-8401141a38d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.041325] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.042067] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276649, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.095214] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276650, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.197706] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276652, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.200636] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276651, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071926} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.200954] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.201707] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5889757f-5db2-4579-85a9-22ee4455ab65 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.223781] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe/71cd9036-0f99-4e30-aad6-ceb4f15d4ffe.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.224065] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0289c6b-e6e0-40cb-9b48-d66eea27af12 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.243686] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 854.243686] env[61867]: value = "task-1276655" [ 854.243686] env[61867]: _type = "Task" [ 854.243686] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.251448] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276655, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.399124] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276654, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.404695] env[61867]: DEBUG nova.scheduler.client.report [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.419967] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7a2f4c7a-00dc-4ac2-a474-94b556ea6cf7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.431493] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ab93f8a-a59a-4051-966c-f488a6a220eb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.468817] env[61867]: DEBUG nova.compute.manager [req-1703cfa2-9eb3-41d6-bb68-35415d88e379 req-74cee087-b243-4495-9216-f40997d387ea service nova] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Detach interface failed, port_id=104495f6-b976-4f53-9959-d5193f833ae8, reason: Instance 9aab8852-addb-49e6-a59b-fa9bffc7733b could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 854.484799] env[61867]: DEBUG nova.compute.manager [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 854.503277] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5298158b-22df-d1f3-7e39-8401141a38d6, 'name': SearchDatastore_Task, 'duration_secs': 0.016374} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.503752] env[61867]: DEBUG oslo_concurrency.lockutils [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.504078] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 854.505138] env[61867]: DEBUG oslo_concurrency.lockutils [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.505138] env[61867]: DEBUG oslo_concurrency.lockutils [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.505138] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.507510] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e676bbf-9e44-4bf7-826e-5c0bc095cda7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.517556] env[61867]: DEBUG nova.virt.hardware [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 854.517879] env[61867]: DEBUG nova.virt.hardware [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 854.518158] env[61867]: DEBUG nova.virt.hardware [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 854.518469] env[61867]: DEBUG nova.virt.hardware [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 854.518607] env[61867]: DEBUG nova.virt.hardware [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 854.518823] env[61867]: DEBUG nova.virt.hardware [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 854.519046] env[61867]: DEBUG nova.virt.hardware [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 854.519291] env[61867]: DEBUG nova.virt.hardware [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 854.519599] env[61867]: DEBUG nova.virt.hardware [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 854.519728] env[61867]: DEBUG nova.virt.hardware [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 854.521387] env[61867]: DEBUG nova.virt.hardware [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 854.521498] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee1d562e-2087-4973-a6e0-5cd60f0ce7a4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.529082] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.529298] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 854.530965] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7fac6b4-f589-4361-99ce-d1d3ff98cc63 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.544029] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2323b43d-7440-4822-b70e-5090c7ed80b2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.547580] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 854.547580] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529115d5-8369-3d50-a577-4efeaa77c09c" [ 854.547580] env[61867]: _type = "Task" [ 854.547580] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.567814] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276649, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.577253] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529115d5-8369-3d50-a577-4efeaa77c09c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.597369] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276650, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.649215] env[61867]: DEBUG nova.network.neutron [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Updating instance_info_cache with network_info: [{"id": "d8069fac-960f-49cf-a416-036e26060a27", "address": "fa:16:3e:ca:61:fe", "network": {"id": "ab6ec5b9-8193-43c9-a3ce-d63f31e5a001", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1573384141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35aa85105b394369b17ef4ce51a49066", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8069fac-96", "ovs_interfaceid": "d8069fac-960f-49cf-a416-036e26060a27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.699089] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276652, 'name': ReconfigVM_Task, 'duration_secs': 0.788477} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.699089] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 410a2819-ea27-4613-9f2c-279f9e82a4b3/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 854.699089] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7438d549-6dca-4416-ada8-876d49e24c78 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.727092] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b409d796-7d7a-4af0-a978-3d4981e6f1eb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.746141] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 854.746141] env[61867]: value = "task-1276656" [ 854.746141] env[61867]: _type = "Task" [ 854.746141] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.759077] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276655, 'name': ReconfigVM_Task, 'duration_secs': 0.35263} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.762673] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe/71cd9036-0f99-4e30-aad6-ceb4f15d4ffe.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 854.763515] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276656, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.763753] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ddd1781b-a935-4c47-a221-6b0849f496aa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.771826] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 854.771826] env[61867]: value = "task-1276657" [ 854.771826] env[61867]: _type = "Task" [ 854.771826] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.787856] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276657, 'name': Rename_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.900330] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276654, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.914974] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.915205] env[61867]: DEBUG nova.compute.manager [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 854.918258] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.303s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.918615] env[61867]: DEBUG nova.objects.instance [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Lazy-loading 'resources' on Instance uuid ad699b24-d01d-4d7a-815f-c6b10286012d {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 855.048188] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276649, 'name': CloneVM_Task, 'duration_secs': 1.811568} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.048417] env[61867]: INFO nova.virt.vmwareapi.vmops [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Created linked-clone VM from snapshot [ 855.049407] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a6fdc6f-4000-445d-932e-df4ac286ef93 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.061901] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529115d5-8369-3d50-a577-4efeaa77c09c, 'name': SearchDatastore_Task, 'duration_secs': 0.080638} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.065994] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Uploading image 444eb1e0-4f45-40cf-b139-01bd39317edd {{(pid=61867) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 855.068413] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96c3ded4-1457-4c67-aaba-7f5242785cee {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.076187] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 855.076187] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5273d2b7-2176-a11d-556f-bd80a9a39bc6" [ 855.076187] env[61867]: _type = "Task" [ 855.076187] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.086822] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5273d2b7-2176-a11d-556f-bd80a9a39bc6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.089079] env[61867]: DEBUG oslo_vmware.rw_handles [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 855.089079] env[61867]: value = "vm-274356" [ 855.089079] env[61867]: _type = "VirtualMachine" [ 855.089079] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 855.089405] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-9e7a2b9c-957b-473f-90a4-58a14029c25c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.100078] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276650, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.101411] env[61867]: DEBUG oslo_vmware.rw_handles [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lease: (returnval){ [ 855.101411] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d7539c-e4c8-e4b1-27d6-913a17344064" [ 855.101411] env[61867]: _type = "HttpNfcLease" [ 855.101411] env[61867]: } obtained for exporting VM: (result){ [ 855.101411] env[61867]: value = "vm-274356" [ 855.101411] env[61867]: _type = "VirtualMachine" [ 855.101411] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 855.101973] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the lease: (returnval){ [ 855.101973] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d7539c-e4c8-e4b1-27d6-913a17344064" [ 855.101973] env[61867]: _type = "HttpNfcLease" [ 855.101973] env[61867]: } to be ready. {{(pid=61867) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 855.110098] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 855.110098] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d7539c-e4c8-e4b1-27d6-913a17344064" [ 855.110098] env[61867]: _type = "HttpNfcLease" [ 855.110098] env[61867]: } is initializing. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 855.154045] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Releasing lock "refresh_cache-6e41989e-b8fa-4009-af1e-1ce859b329a1" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.155950] env[61867]: DEBUG nova.objects.instance [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lazy-loading 'flavor' on Instance uuid 6e41989e-b8fa-4009-af1e-1ce859b329a1 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 855.161356] env[61867]: DEBUG nova.compute.manager [req-5f482571-aacd-48dd-b9d6-96eff57db4a5 req-259d25d0-8f44-47ec-8d60-4cea1c0e5aac service nova] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Received event network-vif-plugged-a3a61d3e-16b8-476c-8ff4-fb229c392f72 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.161611] env[61867]: DEBUG oslo_concurrency.lockutils [req-5f482571-aacd-48dd-b9d6-96eff57db4a5 req-259d25d0-8f44-47ec-8d60-4cea1c0e5aac service nova] Acquiring lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.161828] env[61867]: DEBUG oslo_concurrency.lockutils [req-5f482571-aacd-48dd-b9d6-96eff57db4a5 req-259d25d0-8f44-47ec-8d60-4cea1c0e5aac service nova] Lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.162131] env[61867]: DEBUG oslo_concurrency.lockutils [req-5f482571-aacd-48dd-b9d6-96eff57db4a5 req-259d25d0-8f44-47ec-8d60-4cea1c0e5aac service nova] Lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.162335] env[61867]: DEBUG nova.compute.manager [req-5f482571-aacd-48dd-b9d6-96eff57db4a5 req-259d25d0-8f44-47ec-8d60-4cea1c0e5aac service nova] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] No waiting events found dispatching network-vif-plugged-a3a61d3e-16b8-476c-8ff4-fb229c392f72 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 855.162513] env[61867]: WARNING nova.compute.manager [req-5f482571-aacd-48dd-b9d6-96eff57db4a5 req-259d25d0-8f44-47ec-8d60-4cea1c0e5aac service nova] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Received unexpected event network-vif-plugged-a3a61d3e-16b8-476c-8ff4-fb229c392f72 for instance with vm_state building and task_state spawning. [ 855.256566] env[61867]: DEBUG nova.network.neutron [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Successfully updated port: a3a61d3e-16b8-476c-8ff4-fb229c392f72 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 855.261998] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276656, 'name': ReconfigVM_Task, 'duration_secs': 0.255255} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.262508] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 855.262705] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0621f87d-9e53-447f-927d-bfeb77f48eae {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.272192] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 855.272192] env[61867]: value = "task-1276659" [ 855.272192] env[61867]: _type = "Task" [ 855.272192] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.285619] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276657, 'name': Rename_Task, 'duration_secs': 0.231285} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.289427] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 855.290081] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276659, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.290332] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2e697a78-41b1-4ffa-bc77-1116ced8363a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.298784] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 855.298784] env[61867]: value = "task-1276660" [ 855.298784] env[61867]: _type = "Task" [ 855.298784] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.310535] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276660, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.401540] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276654, 'name': MoveVirtualDisk_Task} progress is 52%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.422156] env[61867]: DEBUG nova.compute.utils [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 855.426896] env[61867]: DEBUG nova.compute.manager [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 855.427067] env[61867]: DEBUG nova.network.neutron [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 855.468949] env[61867]: DEBUG nova.policy [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'adcd9eb75ecc4eccb335ebfec207a900', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7316e4f263a9432ab2f9f91484d62f58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 855.589948] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5273d2b7-2176-a11d-556f-bd80a9a39bc6, 'name': SearchDatastore_Task, 'duration_secs': 0.087461} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.597466] env[61867]: DEBUG oslo_concurrency.lockutils [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.597895] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 6d2dab88-4165-4952-8019-2eaf3b863115/6d2dab88-4165-4952-8019-2eaf3b863115.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 855.598418] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5da53e8-90d7-4773-8365-e009e0810014 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.610567] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 855.610567] env[61867]: value = "task-1276661" [ 855.610567] env[61867]: _type = "Task" [ 855.610567] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.610831] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276650, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.619104] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 855.619104] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d7539c-e4c8-e4b1-27d6-913a17344064" [ 855.619104] env[61867]: _type = "HttpNfcLease" [ 855.619104] env[61867]: } is ready. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 855.620641] env[61867]: DEBUG oslo_vmware.rw_handles [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 855.620641] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d7539c-e4c8-e4b1-27d6-913a17344064" [ 855.620641] env[61867]: _type = "HttpNfcLease" [ 855.620641] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 855.620641] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f5aab9-5222-44e5-b789-6959cbd3535b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.626117] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276661, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.633421] env[61867]: DEBUG oslo_vmware.rw_handles [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c92dd6-b10a-452c-3fa7-e6e08f556007/disk-0.vmdk from lease info. {{(pid=61867) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 855.633623] env[61867]: DEBUG oslo_vmware.rw_handles [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c92dd6-b10a-452c-3fa7-e6e08f556007/disk-0.vmdk for reading. {{(pid=61867) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 855.699313] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fced5fb-9abf-4341-8cd4-d22b38b8e8e3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.724630] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 855.734365] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed34c9c5-54a2-4d39-b5e4-7e3385826043 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.735032] env[61867]: DEBUG oslo_vmware.api [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 855.735032] env[61867]: value = "task-1276662" [ 855.735032] env[61867]: _type = "Task" [ 855.735032] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.744676] env[61867]: DEBUG oslo_vmware.api [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276662, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.759729] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "refresh_cache-2aa08603-d87f-4734-bdfe-fdd610d54e1f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.759891] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquired lock "refresh_cache-2aa08603-d87f-4734-bdfe-fdd610d54e1f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.760061] env[61867]: DEBUG nova.network.neutron [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 855.770025] env[61867]: DEBUG nova.network.neutron [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Successfully created port: b562b890-5eaf-4813-a9ca-66aea05f2342 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 855.781606] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1e83316e-5cca-4734-a531-f7a7a6e240d0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.789489] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276659, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.809099] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d473ccb9-b4f6-42a6-a6fb-76db4640037c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.815106] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276660, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.825349] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27f3307-38a4-41b0-9a81-1eb68e90bc5c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.861543] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd8c631-cf71-4972-9e7e-ff3ab161404e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.870426] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9037603b-5f24-4dd5-8c6b-59cbdce234ca {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.885418] env[61867]: DEBUG nova.compute.provider_tree [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.899021] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276654, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.927848] env[61867]: DEBUG nova.compute.manager [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 856.100169] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276650, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.125629] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276661, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.246598] env[61867]: DEBUG oslo_vmware.api [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276662, 'name': PowerOffVM_Task, 'duration_secs': 0.440253} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.247113] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 856.253761] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Reconfiguring VM instance instance-00000043 to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 856.254121] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94293c4c-6022-4c95-b200-0efeab12aa65 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.286572] env[61867]: DEBUG oslo_vmware.api [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 856.286572] env[61867]: value = "task-1276663" [ 856.286572] env[61867]: _type = "Task" [ 856.286572] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.290305] env[61867]: DEBUG oslo_vmware.api [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276659, 'name': PowerOnVM_Task, 'duration_secs': 0.731643} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.295121] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 856.297666] env[61867]: DEBUG nova.compute.manager [None req-d042e3d9-be25-435e-a6a2-79f27d427e58 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.298721] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48301a66-fd22-4d55-b5d9-c2d17eb3e5f2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.306287] env[61867]: DEBUG oslo_vmware.api [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276663, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.319704] env[61867]: DEBUG oslo_vmware.api [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276660, 'name': PowerOnVM_Task, 'duration_secs': 0.810618} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.320049] env[61867]: DEBUG nova.network.neutron [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 856.322292] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 856.323273] env[61867]: INFO nova.compute.manager [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Took 8.61 seconds to spawn the instance on the hypervisor. [ 856.323273] env[61867]: DEBUG nova.compute.manager [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.324517] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7fb8e0-8cb8-482a-b269-6e12ad2c2882 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.390077] env[61867]: DEBUG nova.scheduler.client.report [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.408945] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276654, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.528031] env[61867]: DEBUG nova.network.neutron [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Updating instance_info_cache with network_info: [{"id": "a3a61d3e-16b8-476c-8ff4-fb229c392f72", "address": "fa:16:3e:e2:4d:d8", "network": {"id": "fc403caf-421f-4783-9807-5e61c4379a6d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-846804042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24180ce8aca142fb897e29d4853c20c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3a61d3e-16", "ovs_interfaceid": "a3a61d3e-16b8-476c-8ff4-fb229c392f72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.600240] env[61867]: DEBUG oslo_vmware.api [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276650, 'name': PowerOnVM_Task, 'duration_secs': 3.161183} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.600555] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 856.600788] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ae23d5c4-21ec-47f4-9159-b80e094d0e4c tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Updating instance '81e70c36-04ea-450c-9383-53ef069d1c46' progress to 100 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 856.624704] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276661, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.806334] env[61867]: DEBUG oslo_vmware.api [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276663, 'name': ReconfigVM_Task, 'duration_secs': 0.414415} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.806625] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Reconfigured VM instance instance-00000043 to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 856.806944] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.807342] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40a446a7-d527-4c9e-976c-d8998ba9a7f8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.816146] env[61867]: DEBUG oslo_vmware.api [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 856.816146] env[61867]: value = "task-1276664" [ 856.816146] env[61867]: _type = "Task" [ 856.816146] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.829876] env[61867]: DEBUG oslo_vmware.api [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276664, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.851442] env[61867]: INFO nova.compute.manager [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Took 29.97 seconds to build instance. [ 856.896287] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.978s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.905600] env[61867]: DEBUG oslo_concurrency.lockutils [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 7.681s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.907115] env[61867]: DEBUG nova.objects.instance [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61867) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 856.922472] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276654, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.792063} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.923105] env[61867]: INFO nova.virt.vmwareapi.ds_util [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_fc57df4d-cf10-49c0-a1d1-738dd6e7ffda/OSTACK_IMG_fc57df4d-cf10-49c0-a1d1-738dd6e7ffda.vmdk to [datastore2] devstack-image-cache_base/420f5356-3ba1-4114-af3a-5be2a65b82cb/420f5356-3ba1-4114-af3a-5be2a65b82cb.vmdk. [ 856.923490] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Cleaning up location [datastore2] OSTACK_IMG_fc57df4d-cf10-49c0-a1d1-738dd6e7ffda {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 856.923784] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_fc57df4d-cf10-49c0-a1d1-738dd6e7ffda {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 856.924281] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e7371c5-2e96-4b60-8dba-8ca698a2835a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.932020] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 856.932020] env[61867]: value = "task-1276665" [ 856.932020] env[61867]: _type = "Task" [ 856.932020] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.938366] env[61867]: INFO nova.scheduler.client.report [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Deleted allocations for instance ad699b24-d01d-4d7a-815f-c6b10286012d [ 856.945197] env[61867]: DEBUG nova.compute.manager [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 856.953216] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276665, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.980451] env[61867]: DEBUG nova.virt.hardware [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 856.980451] env[61867]: DEBUG nova.virt.hardware [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 856.980451] env[61867]: DEBUG nova.virt.hardware [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.980451] env[61867]: DEBUG nova.virt.hardware [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 856.980944] env[61867]: DEBUG nova.virt.hardware [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.981284] env[61867]: DEBUG nova.virt.hardware [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 856.981623] env[61867]: DEBUG nova.virt.hardware [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 856.981936] env[61867]: DEBUG nova.virt.hardware [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 856.982249] env[61867]: DEBUG nova.virt.hardware [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 856.982544] env[61867]: DEBUG nova.virt.hardware [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 856.982835] env[61867]: DEBUG nova.virt.hardware [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 856.983987] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6375315-3efb-47eb-8bd4-d823897b731b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.994276] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31f35698-9dee-447a-9338-a7037f94ceee {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.034522] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Releasing lock "refresh_cache-2aa08603-d87f-4734-bdfe-fdd610d54e1f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.034522] env[61867]: DEBUG nova.compute.manager [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Instance network_info: |[{"id": "a3a61d3e-16b8-476c-8ff4-fb229c392f72", "address": "fa:16:3e:e2:4d:d8", "network": {"id": "fc403caf-421f-4783-9807-5e61c4379a6d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-846804042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24180ce8aca142fb897e29d4853c20c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3a61d3e-16", "ovs_interfaceid": "a3a61d3e-16b8-476c-8ff4-fb229c392f72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 857.034784] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e2:4d:d8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '47ca1ce6-8148-48d5-bcfe-89e39b73914e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a3a61d3e-16b8-476c-8ff4-fb229c392f72', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 857.041508] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Creating folder: Project (24180ce8aca142fb897e29d4853c20c3). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 857.042122] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fa020713-098d-419a-9993-cf6cbb668237 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.053845] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Created folder: Project (24180ce8aca142fb897e29d4853c20c3) in parent group-v274258. [ 857.054322] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Creating folder: Instances. Parent ref: group-v274357. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 857.054686] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-24d8bbc5-ff36-4e05-86a9-e3ea86feb6f0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.065034] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Created folder: Instances in parent group-v274357. [ 857.065034] env[61867]: DEBUG oslo.service.loopingcall [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 857.065034] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 857.065612] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ffb5467e-fbe7-4382-a968-584f208f0baa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.094878] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 857.094878] env[61867]: value = "task-1276668" [ 857.094878] env[61867]: _type = "Task" [ 857.094878] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.103977] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276668, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.127114] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276661, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.450587} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.128881] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 6d2dab88-4165-4952-8019-2eaf3b863115/6d2dab88-4165-4952-8019-2eaf3b863115.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 857.128881] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 857.128881] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb737141-0338-4c1d-a188-7ce160b912e6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.135324] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 857.135324] env[61867]: value = "task-1276669" [ 857.135324] env[61867]: _type = "Task" [ 857.135324] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.144290] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276669, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.300948] env[61867]: DEBUG nova.compute.manager [req-ce796ad7-1f44-43fe-9091-8032e601f187 req-dac0a6d9-0c05-451b-bea0-05c1ea958e34 service nova] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Received event network-changed-a3a61d3e-16b8-476c-8ff4-fb229c392f72 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.301239] env[61867]: DEBUG nova.compute.manager [req-ce796ad7-1f44-43fe-9091-8032e601f187 req-dac0a6d9-0c05-451b-bea0-05c1ea958e34 service nova] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Refreshing instance network info cache due to event network-changed-a3a61d3e-16b8-476c-8ff4-fb229c392f72. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 857.301520] env[61867]: DEBUG oslo_concurrency.lockutils [req-ce796ad7-1f44-43fe-9091-8032e601f187 req-dac0a6d9-0c05-451b-bea0-05c1ea958e34 service nova] Acquiring lock "refresh_cache-2aa08603-d87f-4734-bdfe-fdd610d54e1f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.301650] env[61867]: DEBUG oslo_concurrency.lockutils [req-ce796ad7-1f44-43fe-9091-8032e601f187 req-dac0a6d9-0c05-451b-bea0-05c1ea958e34 service nova] Acquired lock "refresh_cache-2aa08603-d87f-4734-bdfe-fdd610d54e1f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.301887] env[61867]: DEBUG nova.network.neutron [req-ce796ad7-1f44-43fe-9091-8032e601f187 req-dac0a6d9-0c05-451b-bea0-05c1ea958e34 service nova] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Refreshing network info cache for port a3a61d3e-16b8-476c-8ff4-fb229c392f72 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 857.326932] env[61867]: DEBUG oslo_vmware.api [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276664, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.355430] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af4d5b3e-4149-4e7e-9d4d-d5c518a415d1 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.484s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.380469] env[61867]: DEBUG nova.network.neutron [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Successfully updated port: b562b890-5eaf-4813-a9ca-66aea05f2342 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.443999] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276665, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129397} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.444321] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 857.444496] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Releasing lock "[datastore2] devstack-image-cache_base/420f5356-3ba1-4114-af3a-5be2a65b82cb/420f5356-3ba1-4114-af3a-5be2a65b82cb.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.444745] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/420f5356-3ba1-4114-af3a-5be2a65b82cb/420f5356-3ba1-4114-af3a-5be2a65b82cb.vmdk to [datastore2] e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b/e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 857.445011] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e4d8bafc-f3b4-4423-a70f-266b99106c5a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.451481] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 857.451481] env[61867]: value = "task-1276670" [ 857.451481] env[61867]: _type = "Task" [ 857.451481] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.459900] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fad8b535-f8f4-4600-b995-d94cebd9e5f7 tempest-InstanceActionsNegativeTestJSON-207937689 tempest-InstanceActionsNegativeTestJSON-207937689-project-member] Lock "ad699b24-d01d-4d7a-815f-c6b10286012d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.093s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.464739] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276670, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.605762] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276668, 'name': CreateVM_Task} progress is 25%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.653681] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276669, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063994} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.657048] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 857.661194] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd260dc-4aec-4b3c-996a-402a8139f2fd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.690390] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 6d2dab88-4165-4952-8019-2eaf3b863115/6d2dab88-4165-4952-8019-2eaf3b863115.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 857.690390] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80ba4df0-4d6f-4909-9c46-de7e03713562 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.713308] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 857.713308] env[61867]: value = "task-1276671" [ 857.713308] env[61867]: _type = "Task" [ 857.713308] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.722617] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276671, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.832103] env[61867]: DEBUG oslo_vmware.api [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276664, 'name': PowerOnVM_Task, 'duration_secs': 0.619434} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.832103] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.832103] env[61867]: DEBUG nova.compute.manager [None req-f7a6429b-1792-46f5-815b-ec8584abbbe1 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.832103] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda3e453-6348-4302-9878-c9c5ddd50811 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.883182] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "refresh_cache-9c23a44c-eb72-4194-a3e5-88a8ef54ed24" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.883367] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "refresh_cache-9c23a44c-eb72-4194-a3e5-88a8ef54ed24" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.883532] env[61867]: DEBUG nova.network.neutron [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 857.916877] env[61867]: DEBUG oslo_concurrency.lockutils [None req-721f54d0-7251-464b-887f-658aa63aa92a tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.919329] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.096s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.920617] env[61867]: INFO nova.compute.claims [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 857.968214] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276670, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.972119] env[61867]: DEBUG nova.compute.manager [req-b2619e80-202f-41fd-bddd-bcf2d2d55bdf req-a45a0466-070e-4a0c-8a0a-191e4819356f service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Received event network-changed-cf5eac56-8386-49e1-a2b3-d19e3d8297d6 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.972347] env[61867]: DEBUG nova.compute.manager [req-b2619e80-202f-41fd-bddd-bcf2d2d55bdf req-a45a0466-070e-4a0c-8a0a-191e4819356f service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Refreshing instance network info cache due to event network-changed-cf5eac56-8386-49e1-a2b3-d19e3d8297d6. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 857.972644] env[61867]: DEBUG oslo_concurrency.lockutils [req-b2619e80-202f-41fd-bddd-bcf2d2d55bdf req-a45a0466-070e-4a0c-8a0a-191e4819356f service nova] Acquiring lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.972865] env[61867]: DEBUG oslo_concurrency.lockutils [req-b2619e80-202f-41fd-bddd-bcf2d2d55bdf req-a45a0466-070e-4a0c-8a0a-191e4819356f service nova] Acquired lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.973201] env[61867]: DEBUG nova.network.neutron [req-b2619e80-202f-41fd-bddd-bcf2d2d55bdf req-a45a0466-070e-4a0c-8a0a-191e4819356f service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Refreshing network info cache for port cf5eac56-8386-49e1-a2b3-d19e3d8297d6 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 858.107531] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276668, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.229730] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276671, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.461729] env[61867]: DEBUG nova.network.neutron [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 858.471367] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276670, 'name': CopyVirtualDisk_Task} progress is 32%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.524770] env[61867]: DEBUG nova.network.neutron [req-ce796ad7-1f44-43fe-9091-8032e601f187 req-dac0a6d9-0c05-451b-bea0-05c1ea958e34 service nova] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Updated VIF entry in instance network info cache for port a3a61d3e-16b8-476c-8ff4-fb229c392f72. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 858.525258] env[61867]: DEBUG nova.network.neutron [req-ce796ad7-1f44-43fe-9091-8032e601f187 req-dac0a6d9-0c05-451b-bea0-05c1ea958e34 service nova] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Updating instance_info_cache with network_info: [{"id": "a3a61d3e-16b8-476c-8ff4-fb229c392f72", "address": "fa:16:3e:e2:4d:d8", "network": {"id": "fc403caf-421f-4783-9807-5e61c4379a6d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-846804042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24180ce8aca142fb897e29d4853c20c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3a61d3e-16", "ovs_interfaceid": "a3a61d3e-16b8-476c-8ff4-fb229c392f72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.614740] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276668, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.729940] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276671, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.743198] env[61867]: DEBUG nova.network.neutron [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Updating instance_info_cache with network_info: [{"id": "b562b890-5eaf-4813-a9ca-66aea05f2342", "address": "fa:16:3e:a1:df:0e", "network": {"id": "14cee476-296a-4093-99ba-c2a5ef63f7cf", "bridge": "br-int", "label": "tempest-ServersTestJSON-243183479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7316e4f263a9432ab2f9f91484d62f58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb562b890-5e", "ovs_interfaceid": "b562b890-5eaf-4813-a9ca-66aea05f2342", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.926540] env[61867]: DEBUG nova.network.neutron [req-b2619e80-202f-41fd-bddd-bcf2d2d55bdf req-a45a0466-070e-4a0c-8a0a-191e4819356f service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Updated VIF entry in instance network info cache for port cf5eac56-8386-49e1-a2b3-d19e3d8297d6. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 858.927030] env[61867]: DEBUG nova.network.neutron [req-b2619e80-202f-41fd-bddd-bcf2d2d55bdf req-a45a0466-070e-4a0c-8a0a-191e4819356f service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Updating instance_info_cache with network_info: [{"id": "cf5eac56-8386-49e1-a2b3-d19e3d8297d6", "address": "fa:16:3e:a9:7a:17", "network": {"id": "06608a13-6187-41bb-926a-1fbda4f627a0", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-618264839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e54d4ce9d6c24e8ab40cb77886aac634", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf5eac56-83", "ovs_interfaceid": "cf5eac56-8386-49e1-a2b3-d19e3d8297d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.967536] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276670, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.028195] env[61867]: DEBUG oslo_concurrency.lockutils [req-ce796ad7-1f44-43fe-9091-8032e601f187 req-dac0a6d9-0c05-451b-bea0-05c1ea958e34 service nova] Releasing lock "refresh_cache-2aa08603-d87f-4734-bdfe-fdd610d54e1f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.111132] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276668, 'name': CreateVM_Task, 'duration_secs': 1.929056} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.111423] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 859.112354] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.112578] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.113027] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.113376] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-075f3691-5e3d-4f1c-91e1-b1c815329d09 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.121269] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 859.121269] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b2ccd1-f413-4ca4-4346-5f75e4b89849" [ 859.121269] env[61867]: _type = "Task" [ 859.121269] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.134015] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b2ccd1-f413-4ca4-4346-5f75e4b89849, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.232125] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276671, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.243172] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "refresh_cache-9c23a44c-eb72-4194-a3e5-88a8ef54ed24" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.243596] env[61867]: DEBUG nova.compute.manager [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Instance network_info: |[{"id": "b562b890-5eaf-4813-a9ca-66aea05f2342", "address": "fa:16:3e:a1:df:0e", "network": {"id": "14cee476-296a-4093-99ba-c2a5ef63f7cf", "bridge": "br-int", "label": "tempest-ServersTestJSON-243183479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7316e4f263a9432ab2f9f91484d62f58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb562b890-5e", "ovs_interfaceid": "b562b890-5eaf-4813-a9ca-66aea05f2342", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 859.244194] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a1:df:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '085fb0ff-9285-4f1d-a008-a14da4844357', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b562b890-5eaf-4813-a9ca-66aea05f2342', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 859.253233] env[61867]: DEBUG oslo.service.loopingcall [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.256574] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 859.257292] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fc54e5c5-d962-4e57-9142-5a411abee98c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.276923] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb9c9ea-6774-463d-8dab-3f35e32d1225 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.287832] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae97df71-2181-4875-bd40-cd81c57eacd9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.291481] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 859.291481] env[61867]: value = "task-1276672" [ 859.291481] env[61867]: _type = "Task" [ 859.291481] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.325849] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b2bef01-06e0-461f-aa08-5cc2ed3778e7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.332285] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276672, 'name': CreateVM_Task} progress is 15%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.339878] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc31edaa-eac8-43e0-8b44-c768d73bd09a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.360249] env[61867]: DEBUG nova.compute.provider_tree [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.433170] env[61867]: DEBUG oslo_concurrency.lockutils [req-b2619e80-202f-41fd-bddd-bcf2d2d55bdf req-a45a0466-070e-4a0c-8a0a-191e4819356f service nova] Releasing lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.467020] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276670, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.517079] env[61867]: DEBUG oslo_concurrency.lockutils [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.517356] env[61867]: DEBUG oslo_concurrency.lockutils [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.636357] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b2ccd1-f413-4ca4-4346-5f75e4b89849, 'name': SearchDatastore_Task, 'duration_secs': 0.085917} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.636873] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.637455] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 859.638032] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.638247] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.638568] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.639033] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-abf15977-670a-4fc7-9efd-7152e4af21e5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.659645] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.659974] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 859.660722] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38fe3852-6fbe-44e2-aa91-b02d7deaee24 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.668779] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 859.668779] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d9e7cd-bfb5-d92a-faa7-dacbf94271be" [ 859.668779] env[61867]: _type = "Task" [ 859.668779] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.680358] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d9e7cd-bfb5-d92a-faa7-dacbf94271be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.727638] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276671, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.806608] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276672, 'name': CreateVM_Task} progress is 25%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.864794] env[61867]: DEBUG nova.scheduler.client.report [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.967262] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276670, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.996864] env[61867]: DEBUG nova.network.neutron [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Port 79dddbac-ec4f-4cd9-8436-a3600957e598 binding to destination host cpu-1 is already ACTIVE {{(pid=61867) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 859.998256] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.998256] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquired lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.998256] env[61867]: DEBUG nova.network.neutron [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 860.022437] env[61867]: DEBUG nova.compute.utils [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 860.183152] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d9e7cd-bfb5-d92a-faa7-dacbf94271be, 'name': SearchDatastore_Task, 'duration_secs': 0.085373} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.184274] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acef2656-9960-47dc-a452-10b359276f7f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.190832] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 860.190832] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b42953-8255-8c82-e7ec-bd509c49f889" [ 860.190832] env[61867]: _type = "Task" [ 860.190832] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.199518] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b42953-8255-8c82-e7ec-bd509c49f889, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.225896] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276671, 'name': ReconfigVM_Task, 'duration_secs': 2.199768} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.226696] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 6d2dab88-4165-4952-8019-2eaf3b863115/6d2dab88-4165-4952-8019-2eaf3b863115.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.226931] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65d7a543-3d7c-4b6e-94de-508ac10e82ae {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.235105] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 860.235105] env[61867]: value = "task-1276673" [ 860.235105] env[61867]: _type = "Task" [ 860.235105] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.243843] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276673, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.303443] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276672, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.361201] env[61867]: DEBUG nova.compute.manager [req-65c0893a-07fd-40b0-8529-46c129845c4c req-341224a9-c609-44a5-9741-73ea872f44e1 service nova] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Received event network-vif-plugged-b562b890-5eaf-4813-a9ca-66aea05f2342 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.361313] env[61867]: DEBUG oslo_concurrency.lockutils [req-65c0893a-07fd-40b0-8529-46c129845c4c req-341224a9-c609-44a5-9741-73ea872f44e1 service nova] Acquiring lock "9c23a44c-eb72-4194-a3e5-88a8ef54ed24-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.361530] env[61867]: DEBUG oslo_concurrency.lockutils [req-65c0893a-07fd-40b0-8529-46c129845c4c req-341224a9-c609-44a5-9741-73ea872f44e1 service nova] Lock "9c23a44c-eb72-4194-a3e5-88a8ef54ed24-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.361718] env[61867]: DEBUG oslo_concurrency.lockutils [req-65c0893a-07fd-40b0-8529-46c129845c4c req-341224a9-c609-44a5-9741-73ea872f44e1 service nova] Lock "9c23a44c-eb72-4194-a3e5-88a8ef54ed24-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.361906] env[61867]: DEBUG nova.compute.manager [req-65c0893a-07fd-40b0-8529-46c129845c4c req-341224a9-c609-44a5-9741-73ea872f44e1 service nova] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] No waiting events found dispatching network-vif-plugged-b562b890-5eaf-4813-a9ca-66aea05f2342 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 860.362119] env[61867]: WARNING nova.compute.manager [req-65c0893a-07fd-40b0-8529-46c129845c4c req-341224a9-c609-44a5-9741-73ea872f44e1 service nova] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Received unexpected event network-vif-plugged-b562b890-5eaf-4813-a9ca-66aea05f2342 for instance with vm_state building and task_state spawning. [ 860.362299] env[61867]: DEBUG nova.compute.manager [req-65c0893a-07fd-40b0-8529-46c129845c4c req-341224a9-c609-44a5-9741-73ea872f44e1 service nova] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Received event network-changed-b562b890-5eaf-4813-a9ca-66aea05f2342 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.362460] env[61867]: DEBUG nova.compute.manager [req-65c0893a-07fd-40b0-8529-46c129845c4c req-341224a9-c609-44a5-9741-73ea872f44e1 service nova] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Refreshing instance network info cache due to event network-changed-b562b890-5eaf-4813-a9ca-66aea05f2342. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 860.362699] env[61867]: DEBUG oslo_concurrency.lockutils [req-65c0893a-07fd-40b0-8529-46c129845c4c req-341224a9-c609-44a5-9741-73ea872f44e1 service nova] Acquiring lock "refresh_cache-9c23a44c-eb72-4194-a3e5-88a8ef54ed24" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.362803] env[61867]: DEBUG oslo_concurrency.lockutils [req-65c0893a-07fd-40b0-8529-46c129845c4c req-341224a9-c609-44a5-9741-73ea872f44e1 service nova] Acquired lock "refresh_cache-9c23a44c-eb72-4194-a3e5-88a8ef54ed24" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.362988] env[61867]: DEBUG nova.network.neutron [req-65c0893a-07fd-40b0-8529-46c129845c4c req-341224a9-c609-44a5-9741-73ea872f44e1 service nova] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Refreshing network info cache for port b562b890-5eaf-4813-a9ca-66aea05f2342 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 860.374453] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.374453] env[61867]: DEBUG nova.compute.manager [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 860.378845] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.338s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.379226] env[61867]: DEBUG nova.objects.instance [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lazy-loading 'resources' on Instance uuid 9aab8852-addb-49e6-a59b-fa9bffc7733b {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.464846] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276670, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.579959} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.465240] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/420f5356-3ba1-4114-af3a-5be2a65b82cb/420f5356-3ba1-4114-af3a-5be2a65b82cb.vmdk to [datastore2] e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b/e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 860.466091] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ccb5f53-4fc5-406d-a7fc-aa8c67f9f59f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.489550] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b/e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b.vmdk or device None with type streamOptimized {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.490222] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc151069-2ab1-44d8-9070-d01f8a28c28e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.512661] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 860.512661] env[61867]: value = "task-1276674" [ 860.512661] env[61867]: _type = "Task" [ 860.512661] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.522700] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276674, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.525493] env[61867]: DEBUG oslo_concurrency.lockutils [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.704790] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b42953-8255-8c82-e7ec-bd509c49f889, 'name': SearchDatastore_Task, 'duration_secs': 0.014262} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.705132] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.705427] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 2aa08603-d87f-4734-bdfe-fdd610d54e1f/2aa08603-d87f-4734-bdfe-fdd610d54e1f.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 860.705704] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-79a60a3f-1602-4c11-b426-801494215840 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.714814] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 860.714814] env[61867]: value = "task-1276675" [ 860.714814] env[61867]: _type = "Task" [ 860.714814] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.718167] env[61867]: DEBUG nova.compute.manager [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 860.719273] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557b7a9d-8aaa-4a07-b622-57c729ae944a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.726248] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276675, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.745279] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276673, 'name': Rename_Task, 'duration_secs': 0.335494} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.745580] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 860.745834] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef2483ae-4965-48f6-9d13-1b7ab6c42f8b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.751899] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 860.751899] env[61867]: value = "task-1276676" [ 860.751899] env[61867]: _type = "Task" [ 860.751899] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.759640] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276676, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.803104] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276672, 'name': CreateVM_Task, 'duration_secs': 1.021217} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.803422] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 860.804174] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.804588] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.805022] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 860.805288] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2157512e-93f8-4350-87df-5cddb693db9b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.810523] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 860.810523] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d19176-3576-481c-97b2-589543d93a74" [ 860.810523] env[61867]: _type = "Task" [ 860.810523] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.819397] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d19176-3576-481c-97b2-589543d93a74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.880178] env[61867]: DEBUG nova.compute.utils [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 860.881634] env[61867]: DEBUG nova.compute.manager [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 860.881823] env[61867]: DEBUG nova.network.neutron [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 860.909408] env[61867]: DEBUG nova.network.neutron [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Updating instance_info_cache with network_info: [{"id": "79dddbac-ec4f-4cd9-8436-a3600957e598", "address": "fa:16:3e:b8:10:25", "network": {"id": "2d4ab52c-144f-45a1-9d4f-afd85fc30404", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.63", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "84601e35c8e8487cb78fc16a2536a4c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d47d5e1d-e66d-4f2c-83e6-d5e78c2b767d", "external-id": "nsx-vlan-transportzone-109", "segmentation_id": 109, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79dddbac-ec", "ovs_interfaceid": "79dddbac-ec4f-4cd9-8436-a3600957e598", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.964314] env[61867]: DEBUG nova.policy [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12ed008ec3204102b2b08e61ed24f418', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '47c5b5b783ea445eb1c02fb728b1fff2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 861.025190] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276674, 'name': ReconfigVM_Task, 'duration_secs': 0.306062} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.026053] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Reconfigured VM instance instance-0000004a to attach disk [datastore2] e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b/e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b.vmdk or device None with type streamOptimized {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.026349] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9b63e413-3c3e-4dc5-9755-e4ffe3b2d55b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.033308] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 861.033308] env[61867]: value = "task-1276677" [ 861.033308] env[61867]: _type = "Task" [ 861.033308] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.046255] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276677, 'name': Rename_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.178075] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5d294f7-c1ba-4763-b52e-7cd29be03432 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.187237] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a595cf-e1c7-4c6a-9203-9935b92e86bd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.223520] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcc00262-8bb7-4fa3-802e-16000e009a47 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.231484] env[61867]: INFO nova.compute.manager [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] instance snapshotting [ 861.232091] env[61867]: DEBUG nova.objects.instance [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lazy-loading 'flavor' on Instance uuid abb41c0c-6d0d-4147-a4af-554ab7d9e921 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 861.233416] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276675, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.236749] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83965506-0a3e-4f87-a52b-f69108678d1e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.254027] env[61867]: DEBUG nova.compute.provider_tree [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.258463] env[61867]: DEBUG nova.network.neutron [req-65c0893a-07fd-40b0-8529-46c129845c4c req-341224a9-c609-44a5-9741-73ea872f44e1 service nova] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Updated VIF entry in instance network info cache for port b562b890-5eaf-4813-a9ca-66aea05f2342. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 861.258853] env[61867]: DEBUG nova.network.neutron [req-65c0893a-07fd-40b0-8529-46c129845c4c req-341224a9-c609-44a5-9741-73ea872f44e1 service nova] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Updating instance_info_cache with network_info: [{"id": "b562b890-5eaf-4813-a9ca-66aea05f2342", "address": "fa:16:3e:a1:df:0e", "network": {"id": "14cee476-296a-4093-99ba-c2a5ef63f7cf", "bridge": "br-int", "label": "tempest-ServersTestJSON-243183479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7316e4f263a9432ab2f9f91484d62f58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb562b890-5e", "ovs_interfaceid": "b562b890-5eaf-4813-a9ca-66aea05f2342", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.266218] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276676, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.326605] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d19176-3576-481c-97b2-589543d93a74, 'name': SearchDatastore_Task, 'duration_secs': 0.035944} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.326605] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.326605] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 861.326605] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.327189] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.327189] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 861.327189] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-db8e5200-a615-4bce-91f7-3d48e940f35e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.337142] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 861.337359] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 861.338118] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-812dbaa5-d7bb-4ab0-ba9f-f100228ba7f9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.346874] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 861.346874] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52185313-439e-611d-0bf0-d5714534510a" [ 861.346874] env[61867]: _type = "Task" [ 861.346874] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.354215] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52185313-439e-611d-0bf0-d5714534510a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.389959] env[61867]: DEBUG nova.compute.manager [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 861.412265] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Releasing lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.439270] env[61867]: DEBUG nova.compute.manager [req-b997ed49-8815-477d-8087-d74f6d1821f8 req-c8181a51-0697-4e99-aa0d-bdb40d2fb503 service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Received event network-changed-cf5eac56-8386-49e1-a2b3-d19e3d8297d6 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 861.439270] env[61867]: DEBUG nova.compute.manager [req-b997ed49-8815-477d-8087-d74f6d1821f8 req-c8181a51-0697-4e99-aa0d-bdb40d2fb503 service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Refreshing instance network info cache due to event network-changed-cf5eac56-8386-49e1-a2b3-d19e3d8297d6. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 861.439270] env[61867]: DEBUG oslo_concurrency.lockutils [req-b997ed49-8815-477d-8087-d74f6d1821f8 req-c8181a51-0697-4e99-aa0d-bdb40d2fb503 service nova] Acquiring lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.439270] env[61867]: DEBUG oslo_concurrency.lockutils [req-b997ed49-8815-477d-8087-d74f6d1821f8 req-c8181a51-0697-4e99-aa0d-bdb40d2fb503 service nova] Acquired lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.439270] env[61867]: DEBUG nova.network.neutron [req-b997ed49-8815-477d-8087-d74f6d1821f8 req-c8181a51-0697-4e99-aa0d-bdb40d2fb503 service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Refreshing network info cache for port cf5eac56-8386-49e1-a2b3-d19e3d8297d6 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 861.439476] env[61867]: DEBUG nova.network.neutron [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Successfully created port: 2799f597-62eb-4c48-929e-69844f6e0dea {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 861.545979] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276677, 'name': Rename_Task, 'duration_secs': 0.151596} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.546351] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 861.546638] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b71d86ab-74f8-479d-a454-444c7495d321 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.552993] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 861.552993] env[61867]: value = "task-1276678" [ 861.552993] env[61867]: _type = "Task" [ 861.552993] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.561369] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276678, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.620748] env[61867]: DEBUG oslo_concurrency.lockutils [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.621161] env[61867]: DEBUG oslo_concurrency.lockutils [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.621414] env[61867]: INFO nova.compute.manager [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Attaching volume dd584e8e-0be8-44fa-96ac-c5a78546d257 to /dev/sdb [ 861.658991] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ca8e6d0-b749-48d4-aa4c-e5802b8e878e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.667751] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d056be-f77f-4bfa-92f5-dc141323552f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.685126] env[61867]: DEBUG nova.virt.block_device [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Updating existing volume attachment record: 78a92af2-baf2-4482-a634-305c72915a64 {{(pid=61867) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 861.731034] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276675, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.738146] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ed438c-04c5-46fe-9ec0-1c570d731abb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.758685] env[61867]: DEBUG nova.scheduler.client.report [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.766009] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6c1d315-449e-4e3f-ba11-58aa6bfc16e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.769796] env[61867]: DEBUG oslo_concurrency.lockutils [req-65c0893a-07fd-40b0-8529-46c129845c4c req-341224a9-c609-44a5-9741-73ea872f44e1 service nova] Releasing lock "refresh_cache-9c23a44c-eb72-4194-a3e5-88a8ef54ed24" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.783764] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276676, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.857645] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52185313-439e-611d-0bf0-d5714534510a, 'name': SearchDatastore_Task, 'duration_secs': 0.010303} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.858751] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5165a262-dd4b-49b3-8ee7-e377087a852f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.866581] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 861.866581] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52605c78-8ee0-ee84-4fd4-86c1e963ce49" [ 861.866581] env[61867]: _type = "Task" [ 861.866581] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.878895] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52605c78-8ee0-ee84-4fd4-86c1e963ce49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.918246] env[61867]: DEBUG nova.compute.manager [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61867) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 861.918577] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.071150] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276678, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.232594] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276675, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.314821} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.233397] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 2aa08603-d87f-4734-bdfe-fdd610d54e1f/2aa08603-d87f-4734-bdfe-fdd610d54e1f.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 862.234199] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.234639] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3887015e-54f0-4916-af90-b56ac0a418c5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.241493] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 862.241493] env[61867]: value = "task-1276682" [ 862.241493] env[61867]: _type = "Task" [ 862.241493] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.252148] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276682, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.268522] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276676, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.270610] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.892s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.273194] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.355s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.286206] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Creating Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 862.286546] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d71c4f23-5575-488b-92a3-9d0c3d9d7039 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.294729] env[61867]: DEBUG oslo_vmware.api [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 862.294729] env[61867]: value = "task-1276683" [ 862.294729] env[61867]: _type = "Task" [ 862.294729] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.309035] env[61867]: DEBUG oslo_vmware.api [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276683, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.315227] env[61867]: INFO nova.scheduler.client.report [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleted allocations for instance 9aab8852-addb-49e6-a59b-fa9bffc7733b [ 862.352842] env[61867]: DEBUG nova.network.neutron [req-b997ed49-8815-477d-8087-d74f6d1821f8 req-c8181a51-0697-4e99-aa0d-bdb40d2fb503 service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Updated VIF entry in instance network info cache for port cf5eac56-8386-49e1-a2b3-d19e3d8297d6. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 862.353459] env[61867]: DEBUG nova.network.neutron [req-b997ed49-8815-477d-8087-d74f6d1821f8 req-c8181a51-0697-4e99-aa0d-bdb40d2fb503 service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Updating instance_info_cache with network_info: [{"id": "cf5eac56-8386-49e1-a2b3-d19e3d8297d6", "address": "fa:16:3e:a9:7a:17", "network": {"id": "06608a13-6187-41bb-926a-1fbda4f627a0", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-618264839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e54d4ce9d6c24e8ab40cb77886aac634", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf5eac56-83", "ovs_interfaceid": "cf5eac56-8386-49e1-a2b3-d19e3d8297d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.378884] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52605c78-8ee0-ee84-4fd4-86c1e963ce49, 'name': SearchDatastore_Task, 'duration_secs': 0.053962} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.379250] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.379906] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 9c23a44c-eb72-4194-a3e5-88a8ef54ed24/9c23a44c-eb72-4194-a3e5-88a8ef54ed24.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 862.379906] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-69f47017-c9ba-4f32-bb0b-c78e3d4f4e9d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.387421] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 862.387421] env[61867]: value = "task-1276684" [ 862.387421] env[61867]: _type = "Task" [ 862.387421] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.396442] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276684, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.401749] env[61867]: DEBUG nova.compute.manager [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 862.432876] env[61867]: DEBUG nova.virt.hardware [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 862.433087] env[61867]: DEBUG nova.virt.hardware [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 862.433315] env[61867]: DEBUG nova.virt.hardware [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.433510] env[61867]: DEBUG nova.virt.hardware [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 862.433706] env[61867]: DEBUG nova.virt.hardware [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.433891] env[61867]: DEBUG nova.virt.hardware [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 862.434145] env[61867]: DEBUG nova.virt.hardware [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 862.434365] env[61867]: DEBUG nova.virt.hardware [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 862.434551] env[61867]: DEBUG nova.virt.hardware [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 862.434719] env[61867]: DEBUG nova.virt.hardware [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 862.434990] env[61867]: DEBUG nova.virt.hardware [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 862.435997] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d613a0c0-766f-42be-ad46-43687dca4bd4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.444987] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-477a0954-bbd3-4490-b708-0752bc41594e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.562911] env[61867]: DEBUG oslo_vmware.api [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276678, 'name': PowerOnVM_Task, 'duration_secs': 0.540372} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.563268] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 862.563479] env[61867]: INFO nova.compute.manager [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Took 17.17 seconds to spawn the instance on the hypervisor. [ 862.563693] env[61867]: DEBUG nova.compute.manager [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 862.564447] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-855e2a75-d984-4bfb-8474-8dcc44453a6c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.755628] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276682, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074887} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.755968] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.757636] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8536016-bc6f-4b77-bb3e-5da20e1b72b4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.775653] env[61867]: DEBUG oslo_vmware.api [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276676, 'name': PowerOnVM_Task, 'duration_secs': 1.599846} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.788450] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 862.789149] env[61867]: INFO nova.compute.manager [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Took 12.64 seconds to spawn the instance on the hypervisor. [ 862.789549] env[61867]: DEBUG nova.compute.manager [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 862.790112] env[61867]: DEBUG nova.objects.instance [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lazy-loading 'migration_context' on Instance uuid 81e70c36-04ea-450c-9383-53ef069d1c46 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.802806] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 2aa08603-d87f-4734-bdfe-fdd610d54e1f/2aa08603-d87f-4734-bdfe-fdd610d54e1f.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.804850] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d5d7c23-4dfa-4e55-a729-0e2cc61c04ad {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.810998] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b389979c-9097-4a31-a692-b974d51dac21 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.833817] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04cbbf52-652a-4791-b9f8-89f7422ee969 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "9aab8852-addb-49e6-a59b-fa9bffc7733b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.844s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.846768] env[61867]: DEBUG oslo_vmware.api [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276683, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.847120] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 862.847120] env[61867]: value = "task-1276685" [ 862.847120] env[61867]: _type = "Task" [ 862.847120] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.858998] env[61867]: DEBUG oslo_concurrency.lockutils [req-b997ed49-8815-477d-8087-d74f6d1821f8 req-c8181a51-0697-4e99-aa0d-bdb40d2fb503 service nova] Releasing lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.859659] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276685, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.901399] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276684, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.084726] env[61867]: DEBUG nova.compute.manager [req-7c32d3ce-7047-449a-a697-23e9926f6c22 req-e5ba7055-9b9d-436e-9fff-6c3aad2d356a service nova] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Received event network-vif-plugged-2799f597-62eb-4c48-929e-69844f6e0dea {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 863.085108] env[61867]: DEBUG oslo_concurrency.lockutils [req-7c32d3ce-7047-449a-a697-23e9926f6c22 req-e5ba7055-9b9d-436e-9fff-6c3aad2d356a service nova] Acquiring lock "07fdd75c-34ab-45e5-a98b-aced2caa6cb5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.085374] env[61867]: DEBUG oslo_concurrency.lockutils [req-7c32d3ce-7047-449a-a697-23e9926f6c22 req-e5ba7055-9b9d-436e-9fff-6c3aad2d356a service nova] Lock "07fdd75c-34ab-45e5-a98b-aced2caa6cb5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.085687] env[61867]: DEBUG oslo_concurrency.lockutils [req-7c32d3ce-7047-449a-a697-23e9926f6c22 req-e5ba7055-9b9d-436e-9fff-6c3aad2d356a service nova] Lock "07fdd75c-34ab-45e5-a98b-aced2caa6cb5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.086750] env[61867]: DEBUG nova.compute.manager [req-7c32d3ce-7047-449a-a697-23e9926f6c22 req-e5ba7055-9b9d-436e-9fff-6c3aad2d356a service nova] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] No waiting events found dispatching network-vif-plugged-2799f597-62eb-4c48-929e-69844f6e0dea {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 863.086750] env[61867]: WARNING nova.compute.manager [req-7c32d3ce-7047-449a-a697-23e9926f6c22 req-e5ba7055-9b9d-436e-9fff-6c3aad2d356a service nova] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Received unexpected event network-vif-plugged-2799f597-62eb-4c48-929e-69844f6e0dea for instance with vm_state building and task_state spawning. [ 863.090256] env[61867]: INFO nova.compute.manager [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Took 37.49 seconds to build instance. [ 863.215892] env[61867]: DEBUG nova.network.neutron [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Successfully updated port: 2799f597-62eb-4c48-929e-69844f6e0dea {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 863.320958] env[61867]: DEBUG oslo_vmware.api [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276683, 'name': CreateSnapshot_Task, 'duration_secs': 0.962517} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.321377] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Created Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 863.323665] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce19e945-a26b-4902-991e-1a51cb46cb9e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.361859] env[61867]: INFO nova.compute.manager [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Took 34.64 seconds to build instance. [ 863.370113] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276685, 'name': ReconfigVM_Task, 'duration_secs': 0.527939} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.370612] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 2aa08603-d87f-4734-bdfe-fdd610d54e1f/2aa08603-d87f-4734-bdfe-fdd610d54e1f.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 863.371351] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-77007eff-da41-42b3-b9b5-b6acf4f26fcd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.377814] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 863.377814] env[61867]: value = "task-1276686" [ 863.377814] env[61867]: _type = "Task" [ 863.377814] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.392106] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276686, 'name': Rename_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.403244] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276684, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.566807} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.403495] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 9c23a44c-eb72-4194-a3e5-88a8ef54ed24/9c23a44c-eb72-4194-a3e5-88a8ef54ed24.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 863.403716] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 863.403978] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-05a4a9f0-fc00-42f4-bb0b-c0fa676044d7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.411859] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 863.411859] env[61867]: value = "task-1276687" [ 863.411859] env[61867]: _type = "Task" [ 863.411859] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.422174] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276687, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.594168] env[61867]: DEBUG oslo_concurrency.lockutils [None req-04aa1f03-0843-441c-90a4-2e2ca8568e52 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.007s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.606658] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f605a42-7943-4480-a380-297c5bf4cd9b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.615184] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dabf66e-0015-495e-8d0b-8de625008c5b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.646798] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2975fc66-4326-4fda-a208-ecef1cf3ee30 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.654512] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab80788-0c9e-4ca6-a6f8-7ef53ee6b4ec {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.668401] env[61867]: DEBUG nova.compute.provider_tree [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.718660] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "refresh_cache-07fdd75c-34ab-45e5-a98b-aced2caa6cb5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.718817] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "refresh_cache-07fdd75c-34ab-45e5-a98b-aced2caa6cb5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.722034] env[61867]: DEBUG nova.network.neutron [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 863.845448] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Creating linked-clone VM from snapshot {{(pid=61867) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 863.845709] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ee2fb66d-a3bd-4a06-a193-1fdf946a0024 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.855140] env[61867]: DEBUG oslo_vmware.api [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 863.855140] env[61867]: value = "task-1276688" [ 863.855140] env[61867]: _type = "Task" [ 863.855140] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.865163] env[61867]: DEBUG oslo_vmware.api [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276688, 'name': CloneVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.871878] env[61867]: DEBUG oslo_concurrency.lockutils [None req-83225dca-2409-4895-9849-14f4caa0d485 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "6d2dab88-4165-4952-8019-2eaf3b863115" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.994s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.887022] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276686, 'name': Rename_Task, 'duration_secs': 0.239306} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.887310] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 863.887566] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2fac72e9-63cb-4b78-8dc1-ac650a561374 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.893135] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 863.893135] env[61867]: value = "task-1276689" [ 863.893135] env[61867]: _type = "Task" [ 863.893135] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.901934] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276689, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.921384] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276687, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067121} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.921632] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 863.922473] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c9fa16d-b65e-4528-8464-6bd439c72cc9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.929348] env[61867]: DEBUG nova.compute.manager [req-134ade80-d12b-43b7-beb5-322ed53e38bb req-2682dcbc-e869-4772-adce-2e2eafb6bd89 service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Received event network-changed-cf5eac56-8386-49e1-a2b3-d19e3d8297d6 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 863.929548] env[61867]: DEBUG nova.compute.manager [req-134ade80-d12b-43b7-beb5-322ed53e38bb req-2682dcbc-e869-4772-adce-2e2eafb6bd89 service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Refreshing instance network info cache due to event network-changed-cf5eac56-8386-49e1-a2b3-d19e3d8297d6. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 863.929772] env[61867]: DEBUG oslo_concurrency.lockutils [req-134ade80-d12b-43b7-beb5-322ed53e38bb req-2682dcbc-e869-4772-adce-2e2eafb6bd89 service nova] Acquiring lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.929946] env[61867]: DEBUG oslo_concurrency.lockutils [req-134ade80-d12b-43b7-beb5-322ed53e38bb req-2682dcbc-e869-4772-adce-2e2eafb6bd89 service nova] Acquired lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.930110] env[61867]: DEBUG nova.network.neutron [req-134ade80-d12b-43b7-beb5-322ed53e38bb req-2682dcbc-e869-4772-adce-2e2eafb6bd89 service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Refreshing network info cache for port cf5eac56-8386-49e1-a2b3-d19e3d8297d6 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 863.954403] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 9c23a44c-eb72-4194-a3e5-88a8ef54ed24/9c23a44c-eb72-4194-a3e5-88a8ef54ed24.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.959240] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4cbc1344-87d8-4549-a832-29d09f54bb94 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.979028] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 863.979028] env[61867]: value = "task-1276690" [ 863.979028] env[61867]: _type = "Task" [ 863.979028] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.989885] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276690, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.172668] env[61867]: DEBUG nova.scheduler.client.report [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.256259] env[61867]: DEBUG nova.network.neutron [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 864.370419] env[61867]: DEBUG oslo_vmware.api [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276688, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.405488] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276689, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.426036] env[61867]: DEBUG nova.network.neutron [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Updating instance_info_cache with network_info: [{"id": "2799f597-62eb-4c48-929e-69844f6e0dea", "address": "fa:16:3e:30:c6:19", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2799f597-62", "ovs_interfaceid": "2799f597-62eb-4c48-929e-69844f6e0dea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.493530] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276690, 'name': ReconfigVM_Task, 'duration_secs': 0.403573} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.493793] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 9c23a44c-eb72-4194-a3e5-88a8ef54ed24/9c23a44c-eb72-4194-a3e5-88a8ef54ed24.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.494368] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-113252f2-a198-49d3-830d-fc5ae6d1b62a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.502262] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 864.502262] env[61867]: value = "task-1276692" [ 864.502262] env[61867]: _type = "Task" [ 864.502262] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.512814] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276692, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.571792] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "7479bf91-5aef-4e75-a127-7e82ae15a003" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.572228] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "7479bf91-5aef-4e75-a127-7e82ae15a003" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.656328] env[61867]: DEBUG nova.network.neutron [req-134ade80-d12b-43b7-beb5-322ed53e38bb req-2682dcbc-e869-4772-adce-2e2eafb6bd89 service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Updated VIF entry in instance network info cache for port cf5eac56-8386-49e1-a2b3-d19e3d8297d6. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 864.656871] env[61867]: DEBUG nova.network.neutron [req-134ade80-d12b-43b7-beb5-322ed53e38bb req-2682dcbc-e869-4772-adce-2e2eafb6bd89 service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Updating instance_info_cache with network_info: [{"id": "cf5eac56-8386-49e1-a2b3-d19e3d8297d6", "address": "fa:16:3e:a9:7a:17", "network": {"id": "06608a13-6187-41bb-926a-1fbda4f627a0", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-618264839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e54d4ce9d6c24e8ab40cb77886aac634", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf5eac56-83", "ovs_interfaceid": "cf5eac56-8386-49e1-a2b3-d19e3d8297d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.760093] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.760415] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.760982] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.761278] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.761500] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.764028] env[61867]: INFO nova.compute.manager [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Terminating instance [ 864.766063] env[61867]: DEBUG nova.compute.manager [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 864.766299] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 864.767195] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-239d01f0-0edc-4bde-a0a7-57eedab53860 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.775494] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 864.775771] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-77265044-e8d8-4e2a-a26b-789042802582 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.782267] env[61867]: DEBUG oslo_vmware.api [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 864.782267] env[61867]: value = "task-1276693" [ 864.782267] env[61867]: _type = "Task" [ 864.782267] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.793981] env[61867]: DEBUG oslo_vmware.api [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276693, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.872341] env[61867]: DEBUG oslo_vmware.api [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276688, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.905907] env[61867]: DEBUG oslo_vmware.api [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276689, 'name': PowerOnVM_Task, 'duration_secs': 0.863102} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.906926] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 864.906926] env[61867]: INFO nova.compute.manager [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Took 10.42 seconds to spawn the instance on the hypervisor. [ 864.911019] env[61867]: DEBUG nova.compute.manager [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 864.911019] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08f5e1ec-c2aa-4f30-8954-303118e31569 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.928920] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "refresh_cache-07fdd75c-34ab-45e5-a98b-aced2caa6cb5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.929308] env[61867]: DEBUG nova.compute.manager [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Instance network_info: |[{"id": "2799f597-62eb-4c48-929e-69844f6e0dea", "address": "fa:16:3e:30:c6:19", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2799f597-62", "ovs_interfaceid": "2799f597-62eb-4c48-929e-69844f6e0dea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 864.929857] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:c6:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f2020f39-42c4-4481-85c5-aaf03854b459', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2799f597-62eb-4c48-929e-69844f6e0dea', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 864.941994] env[61867]: DEBUG oslo.service.loopingcall [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.942538] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 864.942882] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-76b3c913-5c77-4838-bc22-16d395dcf604 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.971166] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 864.971166] env[61867]: value = "task-1276694" [ 864.971166] env[61867]: _type = "Task" [ 864.971166] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.987179] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276694, 'name': CreateVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.018707] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276692, 'name': Rename_Task, 'duration_secs': 0.173488} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.019260] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 865.019498] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65182f7e-4316-4762-b129-06a655919124 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.027768] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 865.027768] env[61867]: value = "task-1276695" [ 865.027768] env[61867]: _type = "Task" [ 865.027768] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.037637] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276695, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.075737] env[61867]: DEBUG nova.compute.manager [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 865.159786] env[61867]: DEBUG oslo_concurrency.lockutils [req-134ade80-d12b-43b7-beb5-322ed53e38bb req-2682dcbc-e869-4772-adce-2e2eafb6bd89 service nova] Releasing lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.160144] env[61867]: DEBUG nova.compute.manager [req-134ade80-d12b-43b7-beb5-322ed53e38bb req-2682dcbc-e869-4772-adce-2e2eafb6bd89 service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Received event network-changed-cf5eac56-8386-49e1-a2b3-d19e3d8297d6 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.160329] env[61867]: DEBUG nova.compute.manager [req-134ade80-d12b-43b7-beb5-322ed53e38bb req-2682dcbc-e869-4772-adce-2e2eafb6bd89 service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Refreshing instance network info cache due to event network-changed-cf5eac56-8386-49e1-a2b3-d19e3d8297d6. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 865.160559] env[61867]: DEBUG oslo_concurrency.lockutils [req-134ade80-d12b-43b7-beb5-322ed53e38bb req-2682dcbc-e869-4772-adce-2e2eafb6bd89 service nova] Acquiring lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.160820] env[61867]: DEBUG oslo_concurrency.lockutils [req-134ade80-d12b-43b7-beb5-322ed53e38bb req-2682dcbc-e869-4772-adce-2e2eafb6bd89 service nova] Acquired lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.160907] env[61867]: DEBUG nova.network.neutron [req-134ade80-d12b-43b7-beb5-322ed53e38bb req-2682dcbc-e869-4772-adce-2e2eafb6bd89 service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Refreshing network info cache for port cf5eac56-8386-49e1-a2b3-d19e3d8297d6 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 865.184939] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.912s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.303672] env[61867]: DEBUG oslo_vmware.api [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276693, 'name': PowerOffVM_Task, 'duration_secs': 0.21243} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.303977] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 865.304663] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 865.304663] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4d0c2c2e-87d8-44cb-823f-7bf0bfc280bc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.359383] env[61867]: DEBUG nova.compute.manager [req-99b755be-b9d6-49d4-8949-d4809d846865 req-a6bc0809-0e72-4f52-9efb-8e162123d2bc service nova] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Received event network-changed-2799f597-62eb-4c48-929e-69844f6e0dea {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.359613] env[61867]: DEBUG nova.compute.manager [req-99b755be-b9d6-49d4-8949-d4809d846865 req-a6bc0809-0e72-4f52-9efb-8e162123d2bc service nova] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Refreshing instance network info cache due to event network-changed-2799f597-62eb-4c48-929e-69844f6e0dea. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 865.360149] env[61867]: DEBUG oslo_concurrency.lockutils [req-99b755be-b9d6-49d4-8949-d4809d846865 req-a6bc0809-0e72-4f52-9efb-8e162123d2bc service nova] Acquiring lock "refresh_cache-07fdd75c-34ab-45e5-a98b-aced2caa6cb5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.360233] env[61867]: DEBUG oslo_concurrency.lockutils [req-99b755be-b9d6-49d4-8949-d4809d846865 req-a6bc0809-0e72-4f52-9efb-8e162123d2bc service nova] Acquired lock "refresh_cache-07fdd75c-34ab-45e5-a98b-aced2caa6cb5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.360356] env[61867]: DEBUG nova.network.neutron [req-99b755be-b9d6-49d4-8949-d4809d846865 req-a6bc0809-0e72-4f52-9efb-8e162123d2bc service nova] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Refreshing network info cache for port 2799f597-62eb-4c48-929e-69844f6e0dea {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 865.374174] env[61867]: DEBUG oslo_vmware.api [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276688, 'name': CloneVM_Task, 'duration_secs': 1.308234} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.374538] env[61867]: INFO nova.virt.vmwareapi.vmops [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Created linked-clone VM from snapshot [ 865.376644] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c2eda63-e4a1-4112-99c0-148be6d2717f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.386503] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Uploading image 53630313-ba7e-45a9-88e0-b3fbaaa31ae7 {{(pid=61867) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 865.414767] env[61867]: DEBUG oslo_vmware.rw_handles [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 865.414767] env[61867]: value = "vm-274364" [ 865.414767] env[61867]: _type = "VirtualMachine" [ 865.414767] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 865.415190] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-7a2cfb84-c104-4093-a992-c081787bc14e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.418053] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 865.418437] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 865.418547] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Deleting the datastore file [datastore2] e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 865.419089] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e5428c4a-bb52-412b-914b-7876054b38c6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.428494] env[61867]: DEBUG oslo_vmware.rw_handles [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lease: (returnval){ [ 865.428494] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5264cbfe-b140-d0bf-bbec-fa439ddd9bcb" [ 865.428494] env[61867]: _type = "HttpNfcLease" [ 865.428494] env[61867]: } obtained for exporting VM: (result){ [ 865.428494] env[61867]: value = "vm-274364" [ 865.428494] env[61867]: _type = "VirtualMachine" [ 865.428494] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 865.428892] env[61867]: DEBUG oslo_vmware.api [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the lease: (returnval){ [ 865.428892] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5264cbfe-b140-d0bf-bbec-fa439ddd9bcb" [ 865.428892] env[61867]: _type = "HttpNfcLease" [ 865.428892] env[61867]: } to be ready. {{(pid=61867) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 865.430999] env[61867]: DEBUG oslo_vmware.api [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 865.430999] env[61867]: value = "task-1276698" [ 865.430999] env[61867]: _type = "Task" [ 865.430999] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.434125] env[61867]: INFO nova.compute.manager [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Took 31.07 seconds to build instance. [ 865.450456] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 865.450456] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5264cbfe-b140-d0bf-bbec-fa439ddd9bcb" [ 865.450456] env[61867]: _type = "HttpNfcLease" [ 865.450456] env[61867]: } is ready. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 865.450771] env[61867]: DEBUG oslo_vmware.rw_handles [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 865.450771] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5264cbfe-b140-d0bf-bbec-fa439ddd9bcb" [ 865.450771] env[61867]: _type = "HttpNfcLease" [ 865.450771] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 865.452057] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-913e4e02-99b5-436b-b946-bad34dd86c6f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.459659] env[61867]: DEBUG oslo_vmware.api [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276698, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.465898] env[61867]: DEBUG oslo_vmware.rw_handles [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a1ee0c-7d68-3edc-a73e-1fdc35e07db3/disk-0.vmdk from lease info. {{(pid=61867) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 865.466149] env[61867]: DEBUG oslo_vmware.rw_handles [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a1ee0c-7d68-3edc-a73e-1fdc35e07db3/disk-0.vmdk for reading. {{(pid=61867) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 865.535132] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276694, 'name': CreateVM_Task, 'duration_secs': 0.386313} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.535638] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 865.536541] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.536713] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.537058] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 865.540215] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-845443e2-db60-4f73-bbc4-9bfcaebb2974 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.542210] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276695, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.545247] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 865.545247] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5204dd02-c55a-3ac9-6e3e-7b90d1f59c1a" [ 865.545247] env[61867]: _type = "Task" [ 865.545247] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.553577] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5204dd02-c55a-3ac9-6e3e-7b90d1f59c1a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.601194] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.601595] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.603389] env[61867]: INFO nova.compute.claims [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 865.616926] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-da7d76f4-92c0-410a-8a8d-5eb3a1ddfb86 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.907026] env[61867]: DEBUG nova.network.neutron [req-134ade80-d12b-43b7-beb5-322ed53e38bb req-2682dcbc-e869-4772-adce-2e2eafb6bd89 service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Updated VIF entry in instance network info cache for port cf5eac56-8386-49e1-a2b3-d19e3d8297d6. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 865.907436] env[61867]: DEBUG nova.network.neutron [req-134ade80-d12b-43b7-beb5-322ed53e38bb req-2682dcbc-e869-4772-adce-2e2eafb6bd89 service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Updating instance_info_cache with network_info: [{"id": "cf5eac56-8386-49e1-a2b3-d19e3d8297d6", "address": "fa:16:3e:a9:7a:17", "network": {"id": "06608a13-6187-41bb-926a-1fbda4f627a0", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-618264839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e54d4ce9d6c24e8ab40cb77886aac634", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf5eac56-83", "ovs_interfaceid": "cf5eac56-8386-49e1-a2b3-d19e3d8297d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.945166] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c4c8958d-1b8e-4c17-861e-c97dba572334 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.007s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.950939] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Acquiring lock "410a2819-ea27-4613-9f2c-279f9e82a4b3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.951202] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Lock "410a2819-ea27-4613-9f2c-279f9e82a4b3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.951466] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Acquiring lock "410a2819-ea27-4613-9f2c-279f9e82a4b3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.951744] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Lock "410a2819-ea27-4613-9f2c-279f9e82a4b3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.952017] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Lock "410a2819-ea27-4613-9f2c-279f9e82a4b3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.954806] env[61867]: INFO nova.compute.manager [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Terminating instance [ 865.957220] env[61867]: DEBUG nova.compute.manager [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 865.957220] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 865.957531] env[61867]: DEBUG oslo_vmware.api [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276698, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.418132} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.961635] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b867eab9-22ec-47c0-b05e-193c7006802e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.963447] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 865.963570] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 865.963753] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 865.963986] env[61867]: INFO nova.compute.manager [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Took 1.20 seconds to destroy the instance on the hypervisor. [ 865.964613] env[61867]: DEBUG oslo.service.loopingcall [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.968034] env[61867]: DEBUG nova.compute.manager [-] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 865.968145] env[61867]: DEBUG nova.network.neutron [-] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 865.970641] env[61867]: DEBUG nova.compute.manager [req-90b10e73-82a1-4c1b-8063-ad389245d146 req-5adbbb52-d489-4688-91ef-11d9b01bf7a7 service nova] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Received event network-changed-101c4e9b-0368-4c02-9297-4eb478b6fc2b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.971079] env[61867]: DEBUG nova.compute.manager [req-90b10e73-82a1-4c1b-8063-ad389245d146 req-5adbbb52-d489-4688-91ef-11d9b01bf7a7 service nova] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Refreshing instance network info cache due to event network-changed-101c4e9b-0368-4c02-9297-4eb478b6fc2b. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 865.971430] env[61867]: DEBUG oslo_concurrency.lockutils [req-90b10e73-82a1-4c1b-8063-ad389245d146 req-5adbbb52-d489-4688-91ef-11d9b01bf7a7 service nova] Acquiring lock "refresh_cache-6d2dab88-4165-4952-8019-2eaf3b863115" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.971673] env[61867]: DEBUG oslo_concurrency.lockutils [req-90b10e73-82a1-4c1b-8063-ad389245d146 req-5adbbb52-d489-4688-91ef-11d9b01bf7a7 service nova] Acquired lock "refresh_cache-6d2dab88-4165-4952-8019-2eaf3b863115" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.971987] env[61867]: DEBUG nova.network.neutron [req-90b10e73-82a1-4c1b-8063-ad389245d146 req-5adbbb52-d489-4688-91ef-11d9b01bf7a7 service nova] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Refreshing network info cache for port 101c4e9b-0368-4c02-9297-4eb478b6fc2b {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 865.980424] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 865.981072] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-825f45bc-8a51-44f0-9202-50ad27b647d6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.988662] env[61867]: DEBUG oslo_vmware.api [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 865.988662] env[61867]: value = "task-1276699" [ 865.988662] env[61867]: _type = "Task" [ 865.988662] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.000806] env[61867]: DEBUG oslo_vmware.api [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276699, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.047654] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276695, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.065554] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5204dd02-c55a-3ac9-6e3e-7b90d1f59c1a, 'name': SearchDatastore_Task, 'duration_secs': 0.030335} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.065728] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.066228] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 866.066767] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.070020] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.070020] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 866.070020] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-88bbf934-2dd5-4703-b818-313becb9049d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.077477] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 866.077785] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 866.079635] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-230133fc-4c01-4c96-824b-b0c4b49b7e02 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.086791] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 866.086791] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52faac0a-e0a1-2745-451c-2677d95d882d" [ 866.086791] env[61867]: _type = "Task" [ 866.086791] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.095431] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52faac0a-e0a1-2745-451c-2677d95d882d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.265678] env[61867]: DEBUG nova.network.neutron [req-99b755be-b9d6-49d4-8949-d4809d846865 req-a6bc0809-0e72-4f52-9efb-8e162123d2bc service nova] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Updated VIF entry in instance network info cache for port 2799f597-62eb-4c48-929e-69844f6e0dea. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 866.266644] env[61867]: DEBUG nova.network.neutron [req-99b755be-b9d6-49d4-8949-d4809d846865 req-a6bc0809-0e72-4f52-9efb-8e162123d2bc service nova] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Updating instance_info_cache with network_info: [{"id": "2799f597-62eb-4c48-929e-69844f6e0dea", "address": "fa:16:3e:30:c6:19", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2799f597-62", "ovs_interfaceid": "2799f597-62eb-4c48-929e-69844f6e0dea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.334859] env[61867]: DEBUG oslo_vmware.rw_handles [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c92dd6-b10a-452c-3fa7-e6e08f556007/disk-0.vmdk. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 866.336218] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e01389-d218-4c27-847c-fcab2725a6bf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.343319] env[61867]: DEBUG oslo_vmware.rw_handles [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c92dd6-b10a-452c-3fa7-e6e08f556007/disk-0.vmdk is in state: ready. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 866.343588] env[61867]: ERROR oslo_vmware.rw_handles [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c92dd6-b10a-452c-3fa7-e6e08f556007/disk-0.vmdk due to incomplete transfer. [ 866.343987] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f8dcfc13-68c4-4789-9963-86ee4cfbf227 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.350789] env[61867]: DEBUG oslo_vmware.rw_handles [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c92dd6-b10a-452c-3fa7-e6e08f556007/disk-0.vmdk. {{(pid=61867) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 866.351084] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Uploaded image 444eb1e0-4f45-40cf-b139-01bd39317edd to the Glance image server {{(pid=61867) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 866.353863] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Destroying the VM {{(pid=61867) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 866.354109] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-aa317ef3-243d-4b6b-bb50-28549705c32d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.361221] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 866.361221] env[61867]: value = "task-1276700" [ 866.361221] env[61867]: _type = "Task" [ 866.361221] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.371362] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276700, 'name': Destroy_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.410290] env[61867]: DEBUG oslo_concurrency.lockutils [req-134ade80-d12b-43b7-beb5-322ed53e38bb req-2682dcbc-e869-4772-adce-2e2eafb6bd89 service nova] Releasing lock "refresh_cache-410a2819-ea27-4613-9f2c-279f9e82a4b3" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.501211] env[61867]: DEBUG oslo_vmware.api [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276699, 'name': PowerOffVM_Task, 'duration_secs': 0.293015} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.501781] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 866.502082] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 866.502445] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e34974c7-69c1-4cb4-a972-c91f74bec91e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.539342] env[61867]: DEBUG oslo_vmware.api [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276695, 'name': PowerOnVM_Task, 'duration_secs': 1.347538} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.539733] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 866.539945] env[61867]: INFO nova.compute.manager [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Took 9.59 seconds to spawn the instance on the hypervisor. [ 866.540152] env[61867]: DEBUG nova.compute.manager [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.541095] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8987e2d-ab97-41f8-bad1-27a1ac031162 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.599684] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52faac0a-e0a1-2745-451c-2677d95d882d, 'name': SearchDatastore_Task, 'duration_secs': 0.019405} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.601532] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c6ee2d1-023a-41d1-8cfc-633af52c95ee {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.607974] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 866.607974] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52bfe84d-4b32-39db-351e-bac711be5bf1" [ 866.607974] env[61867]: _type = "Task" [ 866.607974] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.619415] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52bfe84d-4b32-39db-351e-bac711be5bf1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.647124] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 866.647660] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 866.647660] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Deleting the datastore file [datastore1] 410a2819-ea27-4613-9f2c-279f9e82a4b3 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 866.648122] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e26ee9fe-4431-4c66-95e7-a24217c88a2d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.654673] env[61867]: DEBUG oslo_vmware.api [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for the task: (returnval){ [ 866.654673] env[61867]: value = "task-1276702" [ 866.654673] env[61867]: _type = "Task" [ 866.654673] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.663845] env[61867]: DEBUG oslo_vmware.api [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276702, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.723727] env[61867]: DEBUG nova.network.neutron [req-90b10e73-82a1-4c1b-8063-ad389245d146 req-5adbbb52-d489-4688-91ef-11d9b01bf7a7 service nova] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Updated VIF entry in instance network info cache for port 101c4e9b-0368-4c02-9297-4eb478b6fc2b. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 866.724090] env[61867]: DEBUG nova.network.neutron [req-90b10e73-82a1-4c1b-8063-ad389245d146 req-5adbbb52-d489-4688-91ef-11d9b01bf7a7 service nova] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Updating instance_info_cache with network_info: [{"id": "101c4e9b-0368-4c02-9297-4eb478b6fc2b", "address": "fa:16:3e:50:da:3d", "network": {"id": "9c6e0d78-8bbe-46b4-9233-94f991e940e8", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1003528124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05dd2be830f9410fbb90415cc13ff6bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap101c4e9b-03", "ovs_interfaceid": "101c4e9b-0368-4c02-9297-4eb478b6fc2b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.726184] env[61867]: INFO nova.compute.manager [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Swapping old allocation on dict_keys(['25720271-a549-4916-abe3-e5ed9b765889']) held by migration 304235dc-4d01-4bfd-af03-d58b4a8ac04a for instance [ 866.746104] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Volume attach. Driver type: vmdk {{(pid=61867) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 866.746483] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274362', 'volume_id': 'dd584e8e-0be8-44fa-96ac-c5a78546d257', 'name': 'volume-dd584e8e-0be8-44fa-96ac-c5a78546d257', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '71cd9036-0f99-4e30-aad6-ceb4f15d4ffe', 'attached_at': '', 'detached_at': '', 'volume_id': 'dd584e8e-0be8-44fa-96ac-c5a78546d257', 'serial': 'dd584e8e-0be8-44fa-96ac-c5a78546d257'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 866.747862] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4dd0d60-0410-41ef-8186-e2f3aaeab84f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.766694] env[61867]: DEBUG nova.scheduler.client.report [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Overwriting current allocation {'allocations': {'25720271-a549-4916-abe3-e5ed9b765889': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 91}}, 'project_id': '2b64ff57036a492eb2555f33e0af664e', 'user_id': '1e18b96b07e54a1d8ed642b02f11e3a4', 'consumer_generation': 1} on consumer 81e70c36-04ea-450c-9383-53ef069d1c46 {{(pid=61867) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 866.772063] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa7f12e0-0731-4b03-ba63-3fb591c02c34 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.775604] env[61867]: DEBUG oslo_concurrency.lockutils [req-99b755be-b9d6-49d4-8949-d4809d846865 req-a6bc0809-0e72-4f52-9efb-8e162123d2bc service nova] Releasing lock "refresh_cache-07fdd75c-34ab-45e5-a98b-aced2caa6cb5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.805153] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] volume-dd584e8e-0be8-44fa-96ac-c5a78546d257/volume-dd584e8e-0be8-44fa-96ac-c5a78546d257.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.809242] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb2a8462-1e8a-4a92-be6e-fac509794191 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.830651] env[61867]: DEBUG oslo_vmware.api [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 866.830651] env[61867]: value = "task-1276703" [ 866.830651] env[61867]: _type = "Task" [ 866.830651] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.849023] env[61867]: DEBUG oslo_vmware.api [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276703, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.871031] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276700, 'name': Destroy_Task} progress is 33%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.889651] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.889846] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquired lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.890046] env[61867]: DEBUG nova.network.neutron [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 866.902877] env[61867]: DEBUG nova.network.neutron [-] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.986784] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2677d157-1e4f-4395-abee-26e3d80879b5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.994939] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00cf8294-d248-4ed0-9960-efc5c99ad6f0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.031844] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9139770-0c39-4269-b228-98bae9376d07 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.040872] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2288db10-a810-409e-b89e-795d2572b441 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.067029] env[61867]: DEBUG nova.compute.provider_tree [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.067029] env[61867]: INFO nova.compute.manager [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Took 31.96 seconds to build instance. [ 867.120248] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52bfe84d-4b32-39db-351e-bac711be5bf1, 'name': SearchDatastore_Task, 'duration_secs': 0.014703} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.120559] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.120908] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 07fdd75c-34ab-45e5-a98b-aced2caa6cb5/07fdd75c-34ab-45e5-a98b-aced2caa6cb5.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 867.121313] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5f0d86eb-95a1-471a-9e43-76534594647c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.127845] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 867.127845] env[61867]: value = "task-1276704" [ 867.127845] env[61867]: _type = "Task" [ 867.127845] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.136118] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276704, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.166092] env[61867]: DEBUG oslo_vmware.api [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Task: {'id': task-1276702, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.41104} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.166959] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.166959] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 867.167198] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 867.167372] env[61867]: INFO nova.compute.manager [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Took 1.21 seconds to destroy the instance on the hypervisor. [ 867.167704] env[61867]: DEBUG oslo.service.loopingcall [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.167865] env[61867]: DEBUG nova.compute.manager [-] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 867.168062] env[61867]: DEBUG nova.network.neutron [-] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 867.229364] env[61867]: DEBUG oslo_concurrency.lockutils [req-90b10e73-82a1-4c1b-8063-ad389245d146 req-5adbbb52-d489-4688-91ef-11d9b01bf7a7 service nova] Releasing lock "refresh_cache-6d2dab88-4165-4952-8019-2eaf3b863115" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.341500] env[61867]: DEBUG oslo_vmware.api [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276703, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.374497] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276700, 'name': Destroy_Task, 'duration_secs': 0.624648} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.374997] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Destroyed the VM [ 867.375559] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Deleting Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 867.376107] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-41794416-0c24-47c8-a5b6-c97bb776d535 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.384728] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 867.384728] env[61867]: value = "task-1276705" [ 867.384728] env[61867]: _type = "Task" [ 867.384728] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.399463] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276705, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.402366] env[61867]: DEBUG nova.compute.manager [req-49f7d151-87e2-4080-ac21-b82ffbd16a87 req-f8d81a8f-1595-420c-8321-384fdd2fb0f2 service nova] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Received event network-vif-deleted-8fe6f5be-9448-4542-a731-fc42b5ab06ca {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.409340] env[61867]: INFO nova.compute.manager [-] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Took 1.44 seconds to deallocate network for instance. [ 867.570028] env[61867]: DEBUG nova.scheduler.client.report [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 867.574158] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b999de9-f221-400a-9b08-4b4b5ca043b0 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "9c23a44c-eb72-4194-a3e5-88a8ef54ed24" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.570s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.641289] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276704, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.479732] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.482924] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.881s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.483577] env[61867]: DEBUG nova.compute.manager [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 868.488158] env[61867]: DEBUG nova.network.neutron [-] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.497926] env[61867]: DEBUG nova.compute.manager [req-c4f1944a-1fce-49a5-8ec4-0878a34c4055 req-5c52c34b-5040-43d0-b4eb-d9f61bcc91dd service nova] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Received event network-changed-a3a61d3e-16b8-476c-8ff4-fb229c392f72 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.499458] env[61867]: DEBUG nova.compute.manager [req-c4f1944a-1fce-49a5-8ec4-0878a34c4055 req-5c52c34b-5040-43d0-b4eb-d9f61bcc91dd service nova] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Refreshing instance network info cache due to event network-changed-a3a61d3e-16b8-476c-8ff4-fb229c392f72. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 868.499458] env[61867]: DEBUG oslo_concurrency.lockutils [req-c4f1944a-1fce-49a5-8ec4-0878a34c4055 req-5c52c34b-5040-43d0-b4eb-d9f61bcc91dd service nova] Acquiring lock "refresh_cache-2aa08603-d87f-4734-bdfe-fdd610d54e1f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.499458] env[61867]: DEBUG oslo_concurrency.lockutils [req-c4f1944a-1fce-49a5-8ec4-0878a34c4055 req-5c52c34b-5040-43d0-b4eb-d9f61bcc91dd service nova] Acquired lock "refresh_cache-2aa08603-d87f-4734-bdfe-fdd610d54e1f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.499458] env[61867]: DEBUG nova.network.neutron [req-c4f1944a-1fce-49a5-8ec4-0878a34c4055 req-5c52c34b-5040-43d0-b4eb-d9f61bcc91dd service nova] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Refreshing network info cache for port a3a61d3e-16b8-476c-8ff4-fb229c392f72 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 868.501962] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.022s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.502133] env[61867]: DEBUG nova.objects.instance [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lazy-loading 'resources' on Instance uuid e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.526379] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276705, 'name': RemoveSnapshot_Task, 'duration_secs': 1.083795} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.527090] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276704, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546018} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.527090] env[61867]: DEBUG oslo_vmware.api [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276703, 'name': ReconfigVM_Task, 'duration_secs': 0.551449} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.527378] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Deleted Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 868.527661] env[61867]: DEBUG nova.compute.manager [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 868.527944] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 07fdd75c-34ab-45e5-a98b-aced2caa6cb5/07fdd75c-34ab-45e5-a98b-aced2caa6cb5.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 868.528165] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 868.528416] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Reconfigured VM instance instance-0000004b to attach disk [datastore2] volume-dd584e8e-0be8-44fa-96ac-c5a78546d257/volume-dd584e8e-0be8-44fa-96ac-c5a78546d257.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 868.536439] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26011584-d845-4b9a-ba8a-f12b38e13be2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.538827] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-12d6d784-05e1-4601-ac05-b0a314a75a88 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.541112] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0df6632-bda2-4239-95dc-784128a3a52c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.560641] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 868.560641] env[61867]: value = "task-1276706" [ 868.560641] env[61867]: _type = "Task" [ 868.560641] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.561425] env[61867]: DEBUG oslo_vmware.api [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 868.561425] env[61867]: value = "task-1276707" [ 868.561425] env[61867]: _type = "Task" [ 868.561425] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.576138] env[61867]: DEBUG oslo_vmware.api [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276707, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.579718] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276706, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.687513] env[61867]: DEBUG nova.network.neutron [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Updating instance_info_cache with network_info: [{"id": "79dddbac-ec4f-4cd9-8436-a3600957e598", "address": "fa:16:3e:b8:10:25", "network": {"id": "2d4ab52c-144f-45a1-9d4f-afd85fc30404", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.63", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "84601e35c8e8487cb78fc16a2536a4c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d47d5e1d-e66d-4f2c-83e6-d5e78c2b767d", "external-id": "nsx-vlan-transportzone-109", "segmentation_id": 109, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79dddbac-ec", "ovs_interfaceid": "79dddbac-ec4f-4cd9-8436-a3600957e598", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.988848] env[61867]: DEBUG nova.compute.utils [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 868.990490] env[61867]: DEBUG nova.compute.manager [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 868.990675] env[61867]: DEBUG nova.network.neutron [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 869.011744] env[61867]: DEBUG oslo_concurrency.lockutils [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "9c23a44c-eb72-4194-a3e5-88a8ef54ed24" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.011744] env[61867]: DEBUG oslo_concurrency.lockutils [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "9c23a44c-eb72-4194-a3e5-88a8ef54ed24" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.011744] env[61867]: DEBUG oslo_concurrency.lockutils [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "9c23a44c-eb72-4194-a3e5-88a8ef54ed24-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.012116] env[61867]: DEBUG oslo_concurrency.lockutils [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "9c23a44c-eb72-4194-a3e5-88a8ef54ed24-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.012116] env[61867]: DEBUG oslo_concurrency.lockutils [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "9c23a44c-eb72-4194-a3e5-88a8ef54ed24-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.013825] env[61867]: INFO nova.compute.manager [-] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Took 1.85 seconds to deallocate network for instance. [ 869.014571] env[61867]: INFO nova.compute.manager [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Terminating instance [ 869.017065] env[61867]: DEBUG nova.compute.manager [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 869.017269] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 869.018368] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3601d00f-af2e-42cb-9884-a899c7dbda4f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.030798] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 869.031141] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7fabd052-bbf1-445b-aae4-aeeeb0cb375c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.040115] env[61867]: DEBUG oslo_vmware.api [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 869.040115] env[61867]: value = "task-1276708" [ 869.040115] env[61867]: _type = "Task" [ 869.040115] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.046187] env[61867]: DEBUG nova.policy [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e5f6dd2ad234411975c32648c366a7d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35aa85105b394369b17ef4ce51a49066', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 869.059472] env[61867]: DEBUG oslo_vmware.api [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276708, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.078128] env[61867]: INFO nova.compute.manager [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Shelve offloading [ 869.079588] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 869.080783] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3755048d-7e77-47c6-9b15-164434777be1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.090689] env[61867]: DEBUG oslo_vmware.api [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276707, 'name': ReconfigVM_Task, 'duration_secs': 0.24681} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.094396] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274362', 'volume_id': 'dd584e8e-0be8-44fa-96ac-c5a78546d257', 'name': 'volume-dd584e8e-0be8-44fa-96ac-c5a78546d257', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '71cd9036-0f99-4e30-aad6-ceb4f15d4ffe', 'attached_at': '', 'detached_at': '', 'volume_id': 'dd584e8e-0be8-44fa-96ac-c5a78546d257', 'serial': 'dd584e8e-0be8-44fa-96ac-c5a78546d257'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 869.096650] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276706, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070666} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.098299] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 869.098549] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 869.098549] env[61867]: value = "task-1276709" [ 869.098549] env[61867]: _type = "Task" [ 869.098549] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.099275] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f10eb61-cd83-40fa-ba91-1225917b6dd6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.127233] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 07fdd75c-34ab-45e5-a98b-aced2caa6cb5/07fdd75c-34ab-45e5-a98b-aced2caa6cb5.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 869.133371] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed3b910d-6b29-4404-aa56-807e3ac3f364 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.153585] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] VM already powered off {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 869.153585] env[61867]: DEBUG nova.compute.manager [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.154608] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b2f7ec-d103-43c5-8416-1d838a7d003b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.161102] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 869.161102] env[61867]: value = "task-1276710" [ 869.161102] env[61867]: _type = "Task" [ 869.161102] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.163164] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.163164] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquired lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.163383] env[61867]: DEBUG nova.network.neutron [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 869.177114] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276710, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.190142] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Releasing lock "refresh_cache-81e70c36-04ea-450c-9383-53ef069d1c46" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.190747] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 869.191131] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a572ca7-3dc1-4d02-b98e-d9e25e39bbff {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.201975] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 869.201975] env[61867]: value = "task-1276711" [ 869.201975] env[61867]: _type = "Task" [ 869.201975] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.211430] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276711, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.382522] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cea7b2f-d482-45c2-b76b-43dcc012a98b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.391840] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd8ba708-9610-4b50-a03e-bc55ad745779 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.426358] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d4bdf4e-d597-4143-b3b6-e454b6642a3c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.435571] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef963b43-9d2f-412d-bd96-b5774c923a36 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.450702] env[61867]: DEBUG nova.compute.provider_tree [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.453859] env[61867]: DEBUG nova.compute.manager [req-d5fe4b23-eff0-4a9c-9208-62f18ea5ce1b req-05ee5a36-7bba-4273-ace3-ce5bb2d08dd1 service nova] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Received event network-vif-deleted-cf5eac56-8386-49e1-a2b3-d19e3d8297d6 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.459336] env[61867]: DEBUG nova.network.neutron [req-c4f1944a-1fce-49a5-8ec4-0878a34c4055 req-5c52c34b-5040-43d0-b4eb-d9f61bcc91dd service nova] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Updated VIF entry in instance network info cache for port a3a61d3e-16b8-476c-8ff4-fb229c392f72. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 869.459739] env[61867]: DEBUG nova.network.neutron [req-c4f1944a-1fce-49a5-8ec4-0878a34c4055 req-5c52c34b-5040-43d0-b4eb-d9f61bcc91dd service nova] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Updating instance_info_cache with network_info: [{"id": "a3a61d3e-16b8-476c-8ff4-fb229c392f72", "address": "fa:16:3e:e2:4d:d8", "network": {"id": "fc403caf-421f-4783-9807-5e61c4379a6d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-846804042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.179", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24180ce8aca142fb897e29d4853c20c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3a61d3e-16", "ovs_interfaceid": "a3a61d3e-16b8-476c-8ff4-fb229c392f72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.493779] env[61867]: DEBUG nova.compute.manager [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 869.526181] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.552652] env[61867]: DEBUG oslo_vmware.api [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276708, 'name': PowerOffVM_Task, 'duration_secs': 0.213392} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.552853] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 869.553139] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 869.553560] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0a677f41-fead-4005-aec2-447f163d06c6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.582180] env[61867]: DEBUG nova.network.neutron [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Successfully created port: 4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 869.629955] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 869.630259] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 869.630474] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleting the datastore file [datastore2] 9c23a44c-eb72-4194-a3e5-88a8ef54ed24 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.630839] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a2bb8849-a050-44f5-a562-ec12ced07150 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.639291] env[61867]: DEBUG oslo_vmware.api [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 869.639291] env[61867]: value = "task-1276713" [ 869.639291] env[61867]: _type = "Task" [ 869.639291] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.650975] env[61867]: DEBUG oslo_vmware.api [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276713, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.674370] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276710, 'name': ReconfigVM_Task, 'duration_secs': 0.409547} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.674669] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 07fdd75c-34ab-45e5-a98b-aced2caa6cb5/07fdd75c-34ab-45e5-a98b-aced2caa6cb5.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 869.675620] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-40430729-3aaa-4546-a546-64e2fcb5342e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.683701] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 869.683701] env[61867]: value = "task-1276714" [ 869.683701] env[61867]: _type = "Task" [ 869.683701] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.693337] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276714, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.716310] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276711, 'name': PowerOffVM_Task, 'duration_secs': 0.231555} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.716707] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 869.717730] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:47:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='f8967eee-6d82-41b2-98fd-bc50fb8c603b',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-993246556',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 869.718032] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 869.718292] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.718540] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 869.718718] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.718954] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 869.719182] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 869.719432] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 869.719656] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 869.719891] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 869.720139] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.729139] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8fb2354-c336-4ec9-9a12-402dd3ed996e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.748486] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 869.748486] env[61867]: value = "task-1276715" [ 869.748486] env[61867]: _type = "Task" [ 869.748486] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.757314] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276715, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.955725] env[61867]: DEBUG nova.scheduler.client.report [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 869.963051] env[61867]: DEBUG oslo_concurrency.lockutils [req-c4f1944a-1fce-49a5-8ec4-0878a34c4055 req-5c52c34b-5040-43d0-b4eb-d9f61bcc91dd service nova] Releasing lock "refresh_cache-2aa08603-d87f-4734-bdfe-fdd610d54e1f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.024444] env[61867]: DEBUG nova.network.neutron [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Updating instance_info_cache with network_info: [{"id": "943c04a5-7eba-46db-ab4f-93f7f84be14e", "address": "fa:16:3e:56:de:39", "network": {"id": "7635ca79-1de7-4ba3-b347-1e0b66d7eb56", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1574772814-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8e9d6e6ceac4520add365429c4eb4d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d0c6fd7-3cc9-4818-9475-8f15900394cc", "external-id": "nsx-vlan-transportzone-317", "segmentation_id": 317, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap943c04a5-7e", "ovs_interfaceid": "943c04a5-7eba-46db-ab4f-93f7f84be14e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.141107] env[61867]: DEBUG nova.objects.instance [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lazy-loading 'flavor' on Instance uuid 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.152452] env[61867]: DEBUG oslo_vmware.api [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276713, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.192705} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.152622] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 870.152804] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 870.153036] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 870.153155] env[61867]: INFO nova.compute.manager [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Took 1.14 seconds to destroy the instance on the hypervisor. [ 870.153401] env[61867]: DEBUG oslo.service.loopingcall [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.154028] env[61867]: DEBUG nova.compute.manager [-] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 870.154028] env[61867]: DEBUG nova.network.neutron [-] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 870.195436] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276714, 'name': Rename_Task, 'duration_secs': 0.190306} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.195436] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 870.195436] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d45be594-3ab3-4c5f-a552-95cbf8a49dff {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.203157] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 870.203157] env[61867]: value = "task-1276716" [ 870.203157] env[61867]: _type = "Task" [ 870.203157] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.213050] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276716, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.261649] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276715, 'name': ReconfigVM_Task, 'duration_secs': 0.167954} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.262853] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a714027-29b9-4732-983d-038421ad0805 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.283597] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:47:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='f8967eee-6d82-41b2-98fd-bc50fb8c603b',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-993246556',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.283810] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.283972] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.284175] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.284328] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.284480] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.284709] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.284885] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.285069] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.285240] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.285413] env[61867]: DEBUG nova.virt.hardware [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.286222] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6dbdfd83-b1a5-4eaa-91e4-fbabf7bcbdb8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.295430] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 870.295430] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52fc6054-2e2c-33d6-fac1-bfef535c8c5f" [ 870.295430] env[61867]: _type = "Task" [ 870.295430] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.305021] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52fc6054-2e2c-33d6-fac1-bfef535c8c5f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.460848] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.959s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.463397] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.937s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.463694] env[61867]: DEBUG nova.objects.instance [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Lazy-loading 'resources' on Instance uuid 410a2819-ea27-4613-9f2c-279f9e82a4b3 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.481841] env[61867]: INFO nova.scheduler.client.report [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Deleted allocations for instance e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b [ 870.503549] env[61867]: DEBUG nova.compute.manager [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 870.525994] env[61867]: DEBUG nova.virt.hardware [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.526324] env[61867]: DEBUG nova.virt.hardware [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.526499] env[61867]: DEBUG nova.virt.hardware [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.526687] env[61867]: DEBUG nova.virt.hardware [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.526841] env[61867]: DEBUG nova.virt.hardware [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.526995] env[61867]: DEBUG nova.virt.hardware [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.527230] env[61867]: DEBUG nova.virt.hardware [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.527394] env[61867]: DEBUG nova.virt.hardware [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.527579] env[61867]: DEBUG nova.virt.hardware [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.527748] env[61867]: DEBUG nova.virt.hardware [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.527925] env[61867]: DEBUG nova.virt.hardware [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.528510] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Releasing lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.530991] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f65f5c96-2159-4b3b-ae78-94880374bca6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.542039] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b613ea9-3d13-412c-813f-70444d3136c5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.650121] env[61867]: DEBUG oslo_concurrency.lockutils [None req-14c72da5-0dce-40c8-a8e1-a1b755bbb8b6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 9.028s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.716826] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276716, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.804518] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52fc6054-2e2c-33d6-fac1-bfef535c8c5f, 'name': SearchDatastore_Task, 'duration_secs': 0.010545} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.809984] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Reconfiguring VM instance instance-0000003f to detach disk 2000 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 870.810353] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98558131-6d91-4806-bc93-e5ce3563d775 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.831659] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 870.831659] env[61867]: value = "task-1276717" [ 870.831659] env[61867]: _type = "Task" [ 870.831659] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.840988] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276717, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.921996] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 870.923067] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe54e38-6ee8-4545-a3f6-cd88e15e74c9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.926453] env[61867]: DEBUG nova.network.neutron [-] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.933633] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 870.933912] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4b229a5-5c24-41c7-9c8a-536a4c945884 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.994933] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4e5753f6-6ed0-4149-82be-f71a481a261c tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.234s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.003619] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 871.003895] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 871.004115] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Deleting the datastore file [datastore1] 1df8427c-e75d-4b60-a92a-b5ba76b67081 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 871.004569] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-682b5761-2c04-447a-a1cf-5633f8120a5a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.016238] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 871.016238] env[61867]: value = "task-1276719" [ 871.016238] env[61867]: _type = "Task" [ 871.016238] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.029720] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276719, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.197198] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f54ea3e-f702-46c1-ae99-503059e59a1a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.208463] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7bac409-0156-412b-880e-b44a45f32f74 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.216671] env[61867]: DEBUG oslo_vmware.api [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276716, 'name': PowerOnVM_Task, 'duration_secs': 0.526224} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.242105] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 871.242359] env[61867]: INFO nova.compute.manager [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Took 8.84 seconds to spawn the instance on the hypervisor. [ 871.242549] env[61867]: DEBUG nova.compute.manager [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 871.243483] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebba853f-46d9-485b-a9f4-de6e5a336a92 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.246315] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c9ecf24-f145-448c-aef2-60d162fd6fd4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.251446] env[61867]: DEBUG oslo_concurrency.lockutils [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.251446] env[61867]: DEBUG oslo_concurrency.lockutils [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.251446] env[61867]: DEBUG oslo_concurrency.lockutils [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.251446] env[61867]: DEBUG oslo_concurrency.lockutils [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.251833] env[61867]: DEBUG oslo_concurrency.lockutils [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.257842] env[61867]: INFO nova.compute.manager [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Terminating instance [ 871.259771] env[61867]: DEBUG nova.network.neutron [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Successfully updated port: 4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 871.262448] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3602c647-8de9-4ab4-9b22-542bcc8ff1d1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.267026] env[61867]: DEBUG nova.compute.manager [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 871.267114] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 871.267380] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70f9ef68-ba0b-4cda-a873-4a18c5707ce0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.280881] env[61867]: DEBUG nova.compute.provider_tree [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.284601] env[61867]: DEBUG oslo_vmware.api [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 871.284601] env[61867]: value = "task-1276720" [ 871.284601] env[61867]: _type = "Task" [ 871.284601] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.297348] env[61867]: DEBUG oslo_vmware.api [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276720, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.343955] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276717, 'name': ReconfigVM_Task, 'duration_secs': 0.264134} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.344332] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Reconfigured VM instance instance-0000003f to detach disk 2000 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 871.345205] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30861b8b-41d0-4a15-9dd9-96b0f4b1ef8e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.382362] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 81e70c36-04ea-450c-9383-53ef069d1c46/81e70c36-04ea-450c-9383-53ef069d1c46.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 871.383310] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d88e3099-ebed-4404-8fad-948955547e4f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.405605] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 871.405605] env[61867]: value = "task-1276721" [ 871.405605] env[61867]: _type = "Task" [ 871.405605] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.415177] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276721, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.429042] env[61867]: INFO nova.compute.manager [-] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Took 1.28 seconds to deallocate network for instance. [ 871.533709] env[61867]: DEBUG nova.compute.manager [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Received event network-vif-deleted-b562b890-5eaf-4813-a9ca-66aea05f2342 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.533984] env[61867]: DEBUG nova.compute.manager [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Received event network-vif-unplugged-943c04a5-7eba-46db-ab4f-93f7f84be14e {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.534226] env[61867]: DEBUG oslo_concurrency.lockutils [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] Acquiring lock "1df8427c-e75d-4b60-a92a-b5ba76b67081-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.534469] env[61867]: DEBUG oslo_concurrency.lockutils [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] Lock "1df8427c-e75d-4b60-a92a-b5ba76b67081-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.534652] env[61867]: DEBUG oslo_concurrency.lockutils [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] Lock "1df8427c-e75d-4b60-a92a-b5ba76b67081-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.534837] env[61867]: DEBUG nova.compute.manager [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] No waiting events found dispatching network-vif-unplugged-943c04a5-7eba-46db-ab4f-93f7f84be14e {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 871.535091] env[61867]: WARNING nova.compute.manager [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Received unexpected event network-vif-unplugged-943c04a5-7eba-46db-ab4f-93f7f84be14e for instance with vm_state shelved and task_state shelving_offloading. [ 871.535361] env[61867]: DEBUG nova.compute.manager [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Received event network-changed-943c04a5-7eba-46db-ab4f-93f7f84be14e {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.535585] env[61867]: DEBUG nova.compute.manager [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Refreshing instance network info cache due to event network-changed-943c04a5-7eba-46db-ab4f-93f7f84be14e. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 871.535834] env[61867]: DEBUG oslo_concurrency.lockutils [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] Acquiring lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.536040] env[61867]: DEBUG oslo_concurrency.lockutils [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] Acquired lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.536257] env[61867]: DEBUG nova.network.neutron [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Refreshing network info cache for port 943c04a5-7eba-46db-ab4f-93f7f84be14e {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 871.538043] env[61867]: DEBUG oslo_vmware.api [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276719, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145002} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.538854] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 871.539173] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 871.539480] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 871.564885] env[61867]: INFO nova.scheduler.client.report [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Deleted allocations for instance 1df8427c-e75d-4b60-a92a-b5ba76b67081 [ 871.767531] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.768057] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.768299] env[61867]: DEBUG nova.network.neutron [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.779757] env[61867]: INFO nova.compute.manager [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Took 19.99 seconds to build instance. [ 871.786393] env[61867]: DEBUG nova.scheduler.client.report [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.802964] env[61867]: DEBUG oslo_vmware.api [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276720, 'name': PowerOffVM_Task, 'duration_secs': 0.292586} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.803400] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 871.803736] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Volume detach. Driver type: vmdk {{(pid=61867) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 871.804104] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274362', 'volume_id': 'dd584e8e-0be8-44fa-96ac-c5a78546d257', 'name': 'volume-dd584e8e-0be8-44fa-96ac-c5a78546d257', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '71cd9036-0f99-4e30-aad6-ceb4f15d4ffe', 'attached_at': '', 'detached_at': '', 'volume_id': 'dd584e8e-0be8-44fa-96ac-c5a78546d257', 'serial': 'dd584e8e-0be8-44fa-96ac-c5a78546d257'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 871.805790] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7e358f-aeb9-4393-83f2-f07043192c0c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.831154] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d3965e-ae42-4e76-a4a4-2cd798c84d72 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.840062] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c208cbbb-d295-473b-98d6-4fb6a3df7cba {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.863154] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ebce3f-9359-4e5b-ab90-526e6158fb91 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.865926] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 871.866233] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 871.881179] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] The volume has not been displaced from its original location: [datastore2] volume-dd584e8e-0be8-44fa-96ac-c5a78546d257/volume-dd584e8e-0be8-44fa-96ac-c5a78546d257.vmdk. No consolidation needed. {{(pid=61867) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 871.886597] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Reconfiguring VM instance instance-0000004b to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 871.886949] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20030e2f-6b8d-4b58-8d07-509a2cf4d590 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.911818] env[61867]: DEBUG oslo_vmware.api [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 871.911818] env[61867]: value = "task-1276722" [ 871.911818] env[61867]: _type = "Task" [ 871.911818] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.921871] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276721, 'name': ReconfigVM_Task, 'duration_secs': 0.326796} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.925402] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 81e70c36-04ea-450c-9383-53ef069d1c46/81e70c36-04ea-450c-9383-53ef069d1c46.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.925788] env[61867]: DEBUG oslo_vmware.api [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276722, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.926539] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89b9576-98c1-4ab2-99d3-93218f55a7f7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.946683] env[61867]: DEBUG oslo_concurrency.lockutils [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.948153] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfdf51b1-dca9-4865-8b81-d463152f97a0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.967513] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b26b3e0-1cb1-4a84-907e-4bd3e32026ff {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.988860] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302fe7d4-4bc2-467c-875f-ff04de0703e7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.997548] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 871.997851] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3693899e-baa1-41b6-8b5f-811fb92776f6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.006443] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 872.006443] env[61867]: value = "task-1276723" [ 872.006443] env[61867]: _type = "Task" [ 872.006443] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.016270] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276723, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.071147] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.211991] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "80a703b3-d692-4023-a73b-dba980a94dff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.212370] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "80a703b3-d692-4023-a73b-dba980a94dff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.213232] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "80a703b3-d692-4023-a73b-dba980a94dff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.213232] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "80a703b3-d692-4023-a73b-dba980a94dff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.213232] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "80a703b3-d692-4023-a73b-dba980a94dff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.216397] env[61867]: INFO nova.compute.manager [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Terminating instance [ 872.217312] env[61867]: DEBUG nova.compute.manager [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 872.217657] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 872.218439] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03810a72-91f0-405c-8fc6-15422ed252e5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.228236] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 872.228914] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8487d24-1f38-49a1-84e2-c64241634748 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.237879] env[61867]: DEBUG oslo_vmware.api [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 872.237879] env[61867]: value = "task-1276724" [ 872.237879] env[61867]: _type = "Task" [ 872.237879] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.250162] env[61867]: DEBUG oslo_vmware.api [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276724, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.281986] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5531776a-378f-4659-967e-316d3d999345 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "07fdd75c-34ab-45e5-a98b-aced2caa6cb5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.502s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.297085] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.833s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.298666] env[61867]: DEBUG oslo_concurrency.lockutils [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.352s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.298918] env[61867]: DEBUG nova.objects.instance [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lazy-loading 'resources' on Instance uuid 9c23a44c-eb72-4194-a3e5-88a8ef54ed24 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 872.323918] env[61867]: INFO nova.scheduler.client.report [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Deleted allocations for instance 410a2819-ea27-4613-9f2c-279f9e82a4b3 [ 872.332286] env[61867]: DEBUG nova.network.neutron [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 872.374991] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 872.375308] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Starting heal instance info cache {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 872.386584] env[61867]: DEBUG nova.network.neutron [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Updated VIF entry in instance network info cache for port 943c04a5-7eba-46db-ab4f-93f7f84be14e. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 872.387131] env[61867]: DEBUG nova.network.neutron [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Updating instance_info_cache with network_info: [{"id": "943c04a5-7eba-46db-ab4f-93f7f84be14e", "address": "fa:16:3e:56:de:39", "network": {"id": "7635ca79-1de7-4ba3-b347-1e0b66d7eb56", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1574772814-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8e9d6e6ceac4520add365429c4eb4d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap943c04a5-7e", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.426628] env[61867]: DEBUG oslo_vmware.api [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276722, 'name': ReconfigVM_Task, 'duration_secs': 0.359987} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.428486] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Reconfigured VM instance instance-0000004b to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 872.435680] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a892dbe1-020c-4644-bebd-3c872936446d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.456117] env[61867]: DEBUG oslo_vmware.api [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 872.456117] env[61867]: value = "task-1276725" [ 872.456117] env[61867]: _type = "Task" [ 872.456117] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.471803] env[61867]: DEBUG oslo_vmware.api [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276725, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.517311] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276723, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.540840] env[61867]: DEBUG nova.network.neutron [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Updating instance_info_cache with network_info: [{"id": "4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1", "address": "fa:16:3e:ce:d6:b5", "network": {"id": "ab6ec5b9-8193-43c9-a3ce-d63f31e5a001", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1573384141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35aa85105b394369b17ef4ce51a49066", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f63bd0e-c1", "ovs_interfaceid": "4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.746987] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "b8e561df-069e-4873-91c3-1bfe5f27de2d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.747263] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "b8e561df-069e-4873-91c3-1bfe5f27de2d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.752497] env[61867]: DEBUG oslo_vmware.api [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276724, 'name': PowerOffVM_Task, 'duration_secs': 0.216387} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.752755] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 872.752936] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 872.753208] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5fa35b27-b3b0-4d09-ae30-cd2d4ca4fa0f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.828798] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 872.829320] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 872.829681] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Deleting the datastore file [datastore2] 80a703b3-d692-4023-a73b-dba980a94dff {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.832616] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-69090288-ae34-483f-a668-9d4d4eaeaa1d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.841272] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dbccbc9c-0c3b-4567-90b6-27885a74fc64 tempest-ServerRescueTestJSONUnderV235-1834517803 tempest-ServerRescueTestJSONUnderV235-1834517803-project-member] Lock "410a2819-ea27-4613-9f2c-279f9e82a4b3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.890s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.844535] env[61867]: DEBUG oslo_vmware.api [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 872.844535] env[61867]: value = "task-1276727" [ 872.844535] env[61867]: _type = "Task" [ 872.844535] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.861030] env[61867]: DEBUG oslo_vmware.api [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276727, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.880152] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Didn't find any instances for network info cache update. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 872.880556] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 872.880803] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 872.881039] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 872.881255] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 872.881751] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 872.882236] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 872.882372] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61867) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 872.882565] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 872.891786] env[61867]: DEBUG oslo_concurrency.lockutils [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] Releasing lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.892024] env[61867]: DEBUG nova.compute.manager [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Received event network-vif-plugged-4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.892231] env[61867]: DEBUG oslo_concurrency.lockutils [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] Acquiring lock "7479bf91-5aef-4e75-a127-7e82ae15a003-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.892435] env[61867]: DEBUG oslo_concurrency.lockutils [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] Lock "7479bf91-5aef-4e75-a127-7e82ae15a003-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.892604] env[61867]: DEBUG oslo_concurrency.lockutils [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] Lock "7479bf91-5aef-4e75-a127-7e82ae15a003-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.892776] env[61867]: DEBUG nova.compute.manager [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] No waiting events found dispatching network-vif-plugged-4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 872.892948] env[61867]: WARNING nova.compute.manager [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Received unexpected event network-vif-plugged-4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1 for instance with vm_state building and task_state spawning. [ 872.893134] env[61867]: DEBUG nova.compute.manager [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Received event network-changed-4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.893300] env[61867]: DEBUG nova.compute.manager [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Refreshing instance network info cache due to event network-changed-4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 872.893465] env[61867]: DEBUG oslo_concurrency.lockutils [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] Acquiring lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.971332] env[61867]: DEBUG oslo_vmware.api [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276725, 'name': ReconfigVM_Task, 'duration_secs': 0.30275} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.974048] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274362', 'volume_id': 'dd584e8e-0be8-44fa-96ac-c5a78546d257', 'name': 'volume-dd584e8e-0be8-44fa-96ac-c5a78546d257', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '71cd9036-0f99-4e30-aad6-ceb4f15d4ffe', 'attached_at': '', 'detached_at': '', 'volume_id': 'dd584e8e-0be8-44fa-96ac-c5a78546d257', 'serial': 'dd584e8e-0be8-44fa-96ac-c5a78546d257'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 872.975342] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 872.975342] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac1c46b-bc37-4167-8515-925d992d9688 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.985260] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 872.985260] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-24098017-8a52-4e82-8558-2b3718a556a9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.017250] env[61867]: DEBUG oslo_vmware.api [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276723, 'name': PowerOnVM_Task, 'duration_secs': 0.537806} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.017546] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 873.045308] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Releasing lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.045676] env[61867]: DEBUG nova.compute.manager [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Instance network_info: |[{"id": "4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1", "address": "fa:16:3e:ce:d6:b5", "network": {"id": "ab6ec5b9-8193-43c9-a3ce-d63f31e5a001", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1573384141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35aa85105b394369b17ef4ce51a49066", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f63bd0e-c1", "ovs_interfaceid": "4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 873.046016] env[61867]: DEBUG oslo_concurrency.lockutils [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] Acquired lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.046220] env[61867]: DEBUG nova.network.neutron [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Refreshing network info cache for port 4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 873.047570] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:d6:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ead20342-9afa-435e-a22b-b4a903457712', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 873.057383] env[61867]: DEBUG oslo.service.loopingcall [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.065124] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 873.065821] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-275e7f6d-1a01-466f-b997-614098dbbca6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.089945] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 873.089945] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 873.089945] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleting the datastore file [datastore2] 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.090734] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccea4285-b379-47c4-ac8a-acbb2060e43c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.094418] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-130738fb-820a-443d-9da2-f4d6e529e9b5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.100103] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 873.100103] env[61867]: value = "task-1276729" [ 873.100103] env[61867]: _type = "Task" [ 873.100103] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.107721] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aebca264-5599-429f-9b01-2136b53e45ca {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.112414] env[61867]: DEBUG oslo_vmware.api [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 873.112414] env[61867]: value = "task-1276730" [ 873.112414] env[61867]: _type = "Task" [ 873.112414] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.153710] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276729, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.157466] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6cf3b58-ffd3-42ad-890f-3835f52bd553 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.164209] env[61867]: DEBUG oslo_vmware.api [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276730, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.171875] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4a12f2e-6148-4bd7-9161-861275b838fa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.193543] env[61867]: DEBUG nova.compute.provider_tree [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.250110] env[61867]: DEBUG nova.compute.manager [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 873.360315] env[61867]: DEBUG oslo_vmware.api [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276727, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.311764} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.360666] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.360867] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 873.361148] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 873.361490] env[61867]: INFO nova.compute.manager [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Took 1.14 seconds to destroy the instance on the hypervisor. [ 873.361669] env[61867]: DEBUG oslo.service.loopingcall [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.361920] env[61867]: DEBUG nova.compute.manager [-] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 873.362067] env[61867]: DEBUG nova.network.neutron [-] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 873.386052] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.582483] env[61867]: DEBUG nova.network.neutron [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Updated VIF entry in instance network info cache for port 4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 873.583046] env[61867]: DEBUG nova.network.neutron [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Updating instance_info_cache with network_info: [{"id": "4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1", "address": "fa:16:3e:ce:d6:b5", "network": {"id": "ab6ec5b9-8193-43c9-a3ce-d63f31e5a001", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1573384141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35aa85105b394369b17ef4ce51a49066", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f63bd0e-c1", "ovs_interfaceid": "4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.613711] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276729, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.623228] env[61867]: DEBUG oslo_vmware.api [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276730, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228639} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.623500] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.623710] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 873.623896] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 873.624089] env[61867]: INFO nova.compute.manager [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Took 2.36 seconds to destroy the instance on the hypervisor. [ 873.624349] env[61867]: DEBUG oslo.service.loopingcall [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.624552] env[61867]: DEBUG nova.compute.manager [-] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 873.624643] env[61867]: DEBUG nova.network.neutron [-] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 873.697528] env[61867]: DEBUG nova.scheduler.client.report [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 873.709022] env[61867]: INFO nova.compute.manager [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Rebuilding instance [ 873.764801] env[61867]: DEBUG nova.compute.manager [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 873.765365] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2033e9be-a287-433e-adf8-84ab03c664d6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.780569] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.039230] env[61867]: INFO nova.compute.manager [None req-4c144004-e6b1-4840-9e6d-a908547167f2 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Updating instance to original state: 'active' [ 874.086534] env[61867]: DEBUG oslo_concurrency.lockutils [req-70e38f60-36cf-40ce-a180-6ef14064e78a req-81b548de-6865-419d-952b-ed4b8c8ff0df service nova] Releasing lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.098011] env[61867]: DEBUG nova.compute.manager [req-25d7bb57-db8a-487f-9729-f2f01852fc76 req-1e198cbd-f9ad-4b5b-aeb6-67114ccab3d6 service nova] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Received event network-vif-deleted-6e8c2dc9-8c61-4169-928a-18c13117f16f {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.098283] env[61867]: INFO nova.compute.manager [req-25d7bb57-db8a-487f-9729-f2f01852fc76 req-1e198cbd-f9ad-4b5b-aeb6-67114ccab3d6 service nova] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Neutron deleted interface 6e8c2dc9-8c61-4169-928a-18c13117f16f; detaching it from the instance and deleting it from the info cache [ 874.098468] env[61867]: DEBUG nova.network.neutron [req-25d7bb57-db8a-487f-9729-f2f01852fc76 req-1e198cbd-f9ad-4b5b-aeb6-67114ccab3d6 service nova] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.114430] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276729, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.212766] env[61867]: DEBUG oslo_concurrency.lockutils [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.914s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.216245] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.145s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.216245] env[61867]: DEBUG nova.objects.instance [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lazy-loading 'resources' on Instance uuid 1df8427c-e75d-4b60-a92a-b5ba76b67081 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.239020] env[61867]: INFO nova.scheduler.client.report [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleted allocations for instance 9c23a44c-eb72-4194-a3e5-88a8ef54ed24 [ 874.277556] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 874.277875] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-018e6f00-2625-4aee-8efd-76706ac4dfdc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.290655] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 874.290655] env[61867]: value = "task-1276731" [ 874.290655] env[61867]: _type = "Task" [ 874.290655] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.301772] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276731, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.364406] env[61867]: DEBUG oslo_concurrency.lockutils [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.364500] env[61867]: DEBUG oslo_concurrency.lockutils [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.516511] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "1df8427c-e75d-4b60-a92a-b5ba76b67081" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.569822] env[61867]: DEBUG nova.network.neutron [-] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.601317] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eccd91e5-5733-47dd-b8da-a2d1f5658f79 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.616647] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276729, 'name': CreateVM_Task, 'duration_secs': 1.467346} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.616930] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 874.619899] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e4baacb-1508-455b-975f-db12ecbc95c6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.631514] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.631716] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.632124] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 874.633286] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50ad331e-8172-4d88-9ffd-d9b5ca5dbccd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.639260] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 874.639260] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520ffd1b-b6ad-2d28-85f0-64e850783ee0" [ 874.639260] env[61867]: _type = "Task" [ 874.639260] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.650119] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520ffd1b-b6ad-2d28-85f0-64e850783ee0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.660223] env[61867]: DEBUG nova.compute.manager [req-25d7bb57-db8a-487f-9729-f2f01852fc76 req-1e198cbd-f9ad-4b5b-aeb6-67114ccab3d6 service nova] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Detach interface failed, port_id=6e8c2dc9-8c61-4169-928a-18c13117f16f, reason: Instance 80a703b3-d692-4023-a73b-dba980a94dff could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 874.718602] env[61867]: DEBUG nova.objects.instance [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lazy-loading 'numa_topology' on Instance uuid 1df8427c-e75d-4b60-a92a-b5ba76b67081 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.746404] env[61867]: DEBUG oslo_concurrency.lockutils [None req-676e1529-5cfb-4e55-891b-eee01fe8b1cf tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "9c23a44c-eb72-4194-a3e5-88a8ef54ed24" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.735s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.802285] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276731, 'name': PowerOffVM_Task, 'duration_secs': 0.243522} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.802772] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 874.803192] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.804123] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-878f0fa5-7438-484a-90d3-05e695998c28 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.813344] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 874.813610] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f29f13e3-2b36-43e6-bb39-ef8cd8c2be23 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.868709] env[61867]: DEBUG nova.compute.utils [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 874.891444] env[61867]: DEBUG nova.network.neutron [-] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.914761] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 874.915305] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 874.915389] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleting the datastore file [datastore1] 07fdd75c-34ab-45e5-a98b-aced2caa6cb5 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 874.917040] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-121b7621-5fcf-467d-a945-562654f05ab3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.924678] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 874.924678] env[61867]: value = "task-1276733" [ 874.924678] env[61867]: _type = "Task" [ 874.924678] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.936624] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276733, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.074563] env[61867]: INFO nova.compute.manager [-] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Took 1.71 seconds to deallocate network for instance. [ 875.154047] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520ffd1b-b6ad-2d28-85f0-64e850783ee0, 'name': SearchDatastore_Task, 'duration_secs': 0.011978} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.154308] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.154495] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 875.154736] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.154889] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.155082] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 875.155370] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a0a489b1-6134-4574-b24e-8607d68e5b0e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.165026] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 875.165275] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 875.166047] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac7ac032-285c-432b-8850-762b8d8d7095 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.172605] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 875.172605] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521c8956-c75d-152c-ce4f-51bea4bf5443" [ 875.172605] env[61867]: _type = "Task" [ 875.172605] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.182594] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521c8956-c75d-152c-ce4f-51bea4bf5443, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.220992] env[61867]: DEBUG nova.objects.base [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Object Instance<1df8427c-e75d-4b60-a92a-b5ba76b67081> lazy-loaded attributes: resources,numa_topology {{(pid=61867) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 875.373066] env[61867]: DEBUG oslo_concurrency.lockutils [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.393722] env[61867]: INFO nova.compute.manager [-] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Took 1.77 seconds to deallocate network for instance. [ 875.435749] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276733, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.221073} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.438325] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.438526] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.438706] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.466199] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e52b906f-ea10-4d3b-829b-3cb0245dcfb5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.474501] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16c3d06-1222-4a8e-83e4-34da01f0580e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.508993] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984bd36f-f45a-4249-81a1-386ee7e2cf1e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.517656] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d1d980-8303-44e7-9466-f3d7ae3497fb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.532732] env[61867]: DEBUG nova.compute.provider_tree [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.582173] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.669493] env[61867]: DEBUG oslo_vmware.rw_handles [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a1ee0c-7d68-3edc-a73e-1fdc35e07db3/disk-0.vmdk. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 875.670384] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7340eba-6405-4c26-a99a-43f7092bdcb9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.680463] env[61867]: DEBUG oslo_vmware.rw_handles [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a1ee0c-7d68-3edc-a73e-1fdc35e07db3/disk-0.vmdk is in state: ready. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 875.680642] env[61867]: ERROR oslo_vmware.rw_handles [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a1ee0c-7d68-3edc-a73e-1fdc35e07db3/disk-0.vmdk due to incomplete transfer. [ 875.683956] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-62e5e572-eb81-46bd-acec-deb33b4cbec4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.685472] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521c8956-c75d-152c-ce4f-51bea4bf5443, 'name': SearchDatastore_Task, 'duration_secs': 0.015142} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.686554] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b609a5e6-fd17-4610-a0e8-f4b001c5bffd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.693189] env[61867]: DEBUG oslo_vmware.rw_handles [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a1ee0c-7d68-3edc-a73e-1fdc35e07db3/disk-0.vmdk. {{(pid=61867) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 875.693433] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Uploaded image 53630313-ba7e-45a9-88e0-b3fbaaa31ae7 to the Glance image server {{(pid=61867) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 875.695227] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Destroying the VM {{(pid=61867) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 875.696400] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-134a7733-650d-49d2-8df5-8772c225a1d0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.697830] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 875.697830] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52c91cb4-8e8f-5997-70a8-f2bafe718718" [ 875.697830] env[61867]: _type = "Task" [ 875.697830] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.703436] env[61867]: DEBUG oslo_vmware.api [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 875.703436] env[61867]: value = "task-1276734" [ 875.703436] env[61867]: _type = "Task" [ 875.703436] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.706588] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52c91cb4-8e8f-5997-70a8-f2bafe718718, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.714168] env[61867]: DEBUG oslo_vmware.api [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276734, 'name': Destroy_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.947126] env[61867]: INFO nova.compute.manager [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Took 0.55 seconds to detach 1 volumes for instance. [ 876.036193] env[61867]: DEBUG nova.scheduler.client.report [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.059927] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "81e70c36-04ea-450c-9383-53ef069d1c46" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.060226] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "81e70c36-04ea-450c-9383-53ef069d1c46" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.060648] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "81e70c36-04ea-450c-9383-53ef069d1c46-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.061063] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "81e70c36-04ea-450c-9383-53ef069d1c46-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.061287] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "81e70c36-04ea-450c-9383-53ef069d1c46-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.063329] env[61867]: INFO nova.compute.manager [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Terminating instance [ 876.065019] env[61867]: DEBUG nova.compute.manager [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 876.065231] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 876.066061] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98464274-1473-4b6d-bbe1-9ab47b9bd24e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.074410] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 876.074671] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-243db464-c329-40c7-9ef9-9e890ca6898e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.081673] env[61867]: DEBUG oslo_vmware.api [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 876.081673] env[61867]: value = "task-1276735" [ 876.081673] env[61867]: _type = "Task" [ 876.081673] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.092278] env[61867]: DEBUG oslo_vmware.api [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276735, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.174282] env[61867]: DEBUG nova.compute.manager [req-ae5068e4-0363-4176-a50b-11d747b310ad req-405035b7-f61c-4e88-bba9-106e3f419499 service nova] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Received event network-vif-deleted-2b465a8c-afc2-4890-a172-d8beb1215f8a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 876.209699] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52c91cb4-8e8f-5997-70a8-f2bafe718718, 'name': SearchDatastore_Task, 'duration_secs': 0.039282} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.212959] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.213246] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 7479bf91-5aef-4e75-a127-7e82ae15a003/7479bf91-5aef-4e75-a127-7e82ae15a003.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 876.213522] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8eefb546-49cc-462c-bee1-afb52de4f999 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.220691] env[61867]: DEBUG oslo_vmware.api [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276734, 'name': Destroy_Task, 'duration_secs': 0.313723} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.221919] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Destroyed the VM [ 876.222208] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Deleting Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 876.222527] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 876.222527] env[61867]: value = "task-1276736" [ 876.222527] env[61867]: _type = "Task" [ 876.222527] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.222740] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-68cb5286-d3b3-426c-b203-d8639b3a9c7d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.233172] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276736, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.234588] env[61867]: DEBUG oslo_vmware.api [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 876.234588] env[61867]: value = "task-1276737" [ 876.234588] env[61867]: _type = "Task" [ 876.234588] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.243969] env[61867]: DEBUG oslo_vmware.api [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276737, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.442330] env[61867]: DEBUG oslo_concurrency.lockutils [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.442634] env[61867]: DEBUG oslo_concurrency.lockutils [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.442881] env[61867]: INFO nova.compute.manager [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Attaching volume b788827a-5734-4c3e-a38d-82845c30c290 to /dev/sdb [ 876.456348] env[61867]: DEBUG oslo_concurrency.lockutils [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.465942] env[61867]: DEBUG oslo_concurrency.lockutils [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "adc2732d-2a10-40ce-bb90-ed0762a36614" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.466318] env[61867]: DEBUG oslo_concurrency.lockutils [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "adc2732d-2a10-40ce-bb90-ed0762a36614" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.481733] env[61867]: DEBUG nova.virt.hardware [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 876.481921] env[61867]: DEBUG nova.virt.hardware [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 876.482097] env[61867]: DEBUG nova.virt.hardware [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.482294] env[61867]: DEBUG nova.virt.hardware [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 876.482447] env[61867]: DEBUG nova.virt.hardware [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.482600] env[61867]: DEBUG nova.virt.hardware [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 876.482819] env[61867]: DEBUG nova.virt.hardware [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 876.482983] env[61867]: DEBUG nova.virt.hardware [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 876.483177] env[61867]: DEBUG nova.virt.hardware [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 876.483391] env[61867]: DEBUG nova.virt.hardware [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 876.483650] env[61867]: DEBUG nova.virt.hardware [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.485071] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841e3632-ea65-4bd7-b347-8dc73e059c85 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.491675] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1819e85e-6a8e-407d-8174-6211fab9277d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.501866] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62aa6bb9-b747-4042-8d13-a3c5f545d913 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.507105] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74943cb4-6ff3-4636-8fac-12b3dc30890f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.523058] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:c6:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f2020f39-42c4-4481-85c5-aaf03854b459', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2799f597-62eb-4c48-929e-69844f6e0dea', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.530607] env[61867]: DEBUG oslo.service.loopingcall [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.535407] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 876.535770] env[61867]: DEBUG nova.virt.block_device [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Updating existing volume attachment record: 35b59762-98a0-479a-964a-4d5be8ec1b57 {{(pid=61867) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 876.538274] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-171ed411-eaf1-4931-a52d-c412dee3b1f3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.554271] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.339s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.557965] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.172s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.557965] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.557965] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61867) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 876.558493] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.778s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.560075] env[61867]: INFO nova.compute.claims [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 876.564390] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5486df8-c854-4974-86f7-b8deb7d90bb1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.570744] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.570744] env[61867]: value = "task-1276738" [ 876.570744] env[61867]: _type = "Task" [ 876.570744] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.577806] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9674f639-7c8a-46b9-9b56-783ddf36a15e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.588462] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276738, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.604530] env[61867]: DEBUG oslo_vmware.api [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276735, 'name': PowerOffVM_Task, 'duration_secs': 0.257732} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.606168] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6813cc12-43bb-400a-ba1a-9638fb9a2533 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.608942] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 876.609681] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 876.610251] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-864f6349-dbc7-42ac-beb0-c86c52883760 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.622427] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad329c38-1b34-4f9c-8621-bb004ca0f4c0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.657753] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180399MB free_disk=141GB free_vcpus=48 pci_devices=None {{(pid=61867) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 876.657939] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.707011] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 876.707296] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 876.707588] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Deleting the datastore file [datastore2] 81e70c36-04ea-450c-9383-53ef069d1c46 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 876.707969] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4f787c78-2e27-473e-8d1e-dd5a94e786ce {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.718151] env[61867]: DEBUG oslo_vmware.api [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for the task: (returnval){ [ 876.718151] env[61867]: value = "task-1276740" [ 876.718151] env[61867]: _type = "Task" [ 876.718151] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.731867] env[61867]: DEBUG oslo_vmware.api [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276740, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.741264] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276736, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.747777] env[61867]: DEBUG oslo_vmware.api [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276737, 'name': RemoveSnapshot_Task} progress is 58%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.969274] env[61867]: DEBUG nova.compute.manager [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 877.062507] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6dd98104-603c-4914-a8e7-c51100003e0f tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "1df8427c-e75d-4b60-a92a-b5ba76b67081" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 26.830s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.064161] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "1df8427c-e75d-4b60-a92a-b5ba76b67081" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 2.547s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.064411] env[61867]: INFO nova.compute.manager [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Unshelving [ 877.083670] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276738, 'name': CreateVM_Task, 'duration_secs': 0.437051} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.083983] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 877.084524] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.084701] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.085057] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 877.085540] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ad33c86-c051-4c01-9392-24d61eaeddd9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.091416] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 877.091416] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523e590a-f7a6-531a-7137-4f700ecbdc9a" [ 877.091416] env[61867]: _type = "Task" [ 877.091416] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.099725] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523e590a-f7a6-531a-7137-4f700ecbdc9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.232079] env[61867]: DEBUG oslo_vmware.api [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Task: {'id': task-1276740, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178931} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.232836] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 877.233097] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 877.233293] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 877.233535] env[61867]: INFO nova.compute.manager [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Took 1.17 seconds to destroy the instance on the hypervisor. [ 877.233814] env[61867]: DEBUG oslo.service.loopingcall [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.234069] env[61867]: DEBUG nova.compute.manager [-] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 877.234182] env[61867]: DEBUG nova.network.neutron [-] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 877.239828] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276736, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544475} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.243182] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 7479bf91-5aef-4e75-a127-7e82ae15a003/7479bf91-5aef-4e75-a127-7e82ae15a003.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 877.243324] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 877.243571] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cbb08ad2-506a-4653-9605-7d90e0cab24e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.251919] env[61867]: DEBUG oslo_vmware.api [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276737, 'name': RemoveSnapshot_Task, 'duration_secs': 0.720059} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.253578] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Deleted Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 877.253886] env[61867]: INFO nova.compute.manager [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Took 15.52 seconds to snapshot the instance on the hypervisor. [ 877.256492] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 877.256492] env[61867]: value = "task-1276744" [ 877.256492] env[61867]: _type = "Task" [ 877.256492] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.265565] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276744, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.492967] env[61867]: DEBUG oslo_concurrency.lockutils [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.602199] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523e590a-f7a6-531a-7137-4f700ecbdc9a, 'name': SearchDatastore_Task, 'duration_secs': 0.009076} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.605511] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.605598] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 877.605796] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.605949] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.606149] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 877.606798] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-711f3d3a-9a50-4676-ac99-fe011fe2378e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.616017] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.616358] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 877.618942] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60b5dcaa-fdbb-4ecd-b824-ead74287adbb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.624699] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 877.624699] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d4d3ea-4fa1-3161-2e11-343b6f46d077" [ 877.624699] env[61867]: _type = "Task" [ 877.624699] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.632327] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d4d3ea-4fa1-3161-2e11-343b6f46d077, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.756715] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b15176ea-8b29-4290-873f-320e3dfa0807 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.775473] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b9e0eb-c5bc-4c92-a667-2767ccfe33ea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.784331] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276744, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067945} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.785149] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 877.786281] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62154655-7e3f-4b0b-9f73-28e19a0164bd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.834260] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0af4bfe-2c2b-4ce1-b3fd-e40c2546db08 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.858450] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 7479bf91-5aef-4e75-a127-7e82ae15a003/7479bf91-5aef-4e75-a127-7e82ae15a003.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 877.860549] env[61867]: DEBUG nova.compute.manager [None req-ce6e726e-6712-4d45-b44a-90c5590f3490 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Found 1 images (rotation: 2) {{(pid=61867) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 877.865018] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8090b2c8-bd9f-4433-97ff-192bb20e301b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.881427] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d65dc61-3b0c-449f-9fd5-1109d79926b7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.887044] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 877.887044] env[61867]: value = "task-1276745" [ 877.887044] env[61867]: _type = "Task" [ 877.887044] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.899294] env[61867]: DEBUG nova.compute.provider_tree [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.905825] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276745, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.997076] env[61867]: DEBUG nova.network.neutron [-] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.089886] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.135958] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d4d3ea-4fa1-3161-2e11-343b6f46d077, 'name': SearchDatastore_Task, 'duration_secs': 0.008871} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.136789] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87e19433-a88d-4f8f-9f94-63292c821422 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.142310] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 878.142310] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52dc006d-5db6-593b-be75-092491360f63" [ 878.142310] env[61867]: _type = "Task" [ 878.142310] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.150569] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52dc006d-5db6-593b-be75-092491360f63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.199900] env[61867]: DEBUG nova.compute.manager [req-0d9026b7-c4c4-41ed-9fc7-32cea80f81fa req-54a7a159-6797-4193-95f6-01ddeb237c63 service nova] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Received event network-vif-deleted-79dddbac-ec4f-4cd9-8436-a3600957e598 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.398609] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276745, 'name': ReconfigVM_Task, 'duration_secs': 0.473631} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.398841] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 7479bf91-5aef-4e75-a127-7e82ae15a003/7479bf91-5aef-4e75-a127-7e82ae15a003.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 878.399480] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-501b8710-bb18-4934-bce2-6a6531e11622 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.402155] env[61867]: DEBUG nova.scheduler.client.report [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 878.414055] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 878.414055] env[61867]: value = "task-1276746" [ 878.414055] env[61867]: _type = "Task" [ 878.414055] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.423423] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276746, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.499800] env[61867]: INFO nova.compute.manager [-] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Took 1.27 seconds to deallocate network for instance. [ 878.652899] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52dc006d-5db6-593b-be75-092491360f63, 'name': SearchDatastore_Task, 'duration_secs': 0.0105} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.653221] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.653509] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 07fdd75c-34ab-45e5-a98b-aced2caa6cb5/07fdd75c-34ab-45e5-a98b-aced2caa6cb5.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 878.653777] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f7370cd1-0821-4023-bef8-6c5fdbb8d499 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.661455] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 878.661455] env[61867]: value = "task-1276747" [ 878.661455] env[61867]: _type = "Task" [ 878.661455] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.669147] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276747, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.770453] env[61867]: DEBUG nova.compute.manager [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.771590] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7bbf30c-1ea7-4c63-b3c3-6d2e96c8907f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.908652] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.909310] env[61867]: DEBUG nova.compute.manager [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 878.912534] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.330s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.912900] env[61867]: DEBUG nova.objects.instance [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lazy-loading 'resources' on Instance uuid 80a703b3-d692-4023-a73b-dba980a94dff {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.927827] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276746, 'name': Rename_Task, 'duration_secs': 0.136723} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.928146] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 878.928411] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a14412f7-b6c4-4723-9958-d515faa69da0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.937881] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 878.937881] env[61867]: value = "task-1276749" [ 878.937881] env[61867]: _type = "Task" [ 878.937881] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.948639] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276749, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.007490] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.172271] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276747, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.434365} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.172653] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 07fdd75c-34ab-45e5-a98b-aced2caa6cb5/07fdd75c-34ab-45e5-a98b-aced2caa6cb5.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 879.172910] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.173194] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-38c38013-69c1-4141-9d2e-99fb479ad418 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.180213] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 879.180213] env[61867]: value = "task-1276750" [ 879.180213] env[61867]: _type = "Task" [ 879.180213] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.189173] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276750, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.283259] env[61867]: INFO nova.compute.manager [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] instance snapshotting [ 879.284078] env[61867]: DEBUG nova.objects.instance [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lazy-loading 'flavor' on Instance uuid abb41c0c-6d0d-4147-a4af-554ab7d9e921 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 879.416340] env[61867]: DEBUG nova.compute.utils [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 879.418098] env[61867]: DEBUG nova.compute.manager [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 879.418734] env[61867]: DEBUG nova.network.neutron [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 879.450874] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276749, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.466909] env[61867]: DEBUG nova.policy [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc358009ebe6495a881034439d00978f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a29c5e9cfdaa4cc88ed300100a308ffa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 879.654749] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-483ccc72-fe41-4ff9-8cfb-4c6dcd5e1d80 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.665110] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-153a12e9-eecc-43bd-a95e-5bcbe27058db {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.702245] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b138bd2c-44b7-4aee-91ca-46888c25c398 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.712924] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a128665b-1111-4381-b243-878dbd41f5d2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.717545] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276750, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.10944} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.717763] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.719242] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87488b07-4577-4383-a3cb-8002b3f389fc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.731358] env[61867]: DEBUG nova.compute.provider_tree [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.753056] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 07fdd75c-34ab-45e5-a98b-aced2caa6cb5/07fdd75c-34ab-45e5-a98b-aced2caa6cb5.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.754717] env[61867]: DEBUG nova.scheduler.client.report [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.757846] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b77cb681-0815-409d-b07f-87f581578924 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.778930] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.865s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.780569] env[61867]: DEBUG oslo_concurrency.lockutils [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.323s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.780569] env[61867]: DEBUG nova.objects.instance [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lazy-loading 'resources' on Instance uuid 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 879.790701] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 879.790701] env[61867]: value = "task-1276751" [ 879.790701] env[61867]: _type = "Task" [ 879.790701] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.795405] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef774132-0ae9-4b9b-8277-7cbb654fc93a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.799862] env[61867]: DEBUG nova.network.neutron [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Successfully created port: 41bb6052-bacd-4d6b-abd6-b5186a2418e7 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 879.823745] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276751, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.825302] env[61867]: INFO nova.scheduler.client.report [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Deleted allocations for instance 80a703b3-d692-4023-a73b-dba980a94dff [ 879.826733] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f7c470-165c-4b43-93a4-ff385be690e2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.922166] env[61867]: DEBUG nova.compute.manager [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 879.949934] env[61867]: DEBUG oslo_vmware.api [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276749, 'name': PowerOnVM_Task, 'duration_secs': 0.760899} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.950209] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 879.950419] env[61867]: INFO nova.compute.manager [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Took 9.45 seconds to spawn the instance on the hypervisor. [ 879.950648] env[61867]: DEBUG nova.compute.manager [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 879.951473] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0997bd6-f607-458e-bdb2-9c109a90b449 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.301420] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276751, 'name': ReconfigVM_Task, 'duration_secs': 0.274612} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.301811] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 07fdd75c-34ab-45e5-a98b-aced2caa6cb5/07fdd75c-34ab-45e5-a98b-aced2caa6cb5.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.302620] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6870b0de-a0ef-497c-bf76-7312d39a2807 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.310388] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 880.310388] env[61867]: value = "task-1276752" [ 880.310388] env[61867]: _type = "Task" [ 880.310388] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.319227] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276752, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.338475] env[61867]: DEBUG oslo_concurrency.lockutils [None req-92df3a3f-2504-41a9-8000-674de1089d62 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "80a703b3-d692-4023-a73b-dba980a94dff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.126s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.340036] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Creating Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 880.340289] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cdc6023d-34d5-4ada-bd7d-a4e83390012a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.348964] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 880.348964] env[61867]: value = "task-1276753" [ 880.348964] env[61867]: _type = "Task" [ 880.348964] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.360031] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276753, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.471514] env[61867]: INFO nova.compute.manager [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Took 14.89 seconds to build instance. [ 880.493577] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af512aa-1c76-4b63-9510-55c3a31732d3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.502529] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b43443-2f1a-400a-b825-b2305b8ba119 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.534021] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a97766d4-2ec8-426c-9660-f7bff5a52bc7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.542660] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc90d05a-33a7-422c-9690-008aa6c6ee32 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.558061] env[61867]: DEBUG nova.compute.provider_tree [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.823152] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276752, 'name': Rename_Task, 'duration_secs': 0.381466} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.823537] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 880.824117] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6dc978af-ce03-4b7b-ac84-8ab0c1860b19 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.832049] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 880.832049] env[61867]: value = "task-1276754" [ 880.832049] env[61867]: _type = "Task" [ 880.832049] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.840310] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276754, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.858164] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276753, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.933688] env[61867]: DEBUG nova.compute.manager [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 880.961159] env[61867]: DEBUG nova.virt.hardware [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 880.961483] env[61867]: DEBUG nova.virt.hardware [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 880.961599] env[61867]: DEBUG nova.virt.hardware [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.961784] env[61867]: DEBUG nova.virt.hardware [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 880.961936] env[61867]: DEBUG nova.virt.hardware [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.962285] env[61867]: DEBUG nova.virt.hardware [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 880.962804] env[61867]: DEBUG nova.virt.hardware [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 880.962804] env[61867]: DEBUG nova.virt.hardware [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 880.962971] env[61867]: DEBUG nova.virt.hardware [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 880.963137] env[61867]: DEBUG nova.virt.hardware [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 880.963420] env[61867]: DEBUG nova.virt.hardware [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 880.964334] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4181fdf9-892e-40de-a6ad-1771289cdcb8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.973324] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e1f962-9740-48ff-b685-64be5f8a7599 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.977384] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e36edbdb-96f1-47cb-8079-959fe3fcf632 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "7479bf91-5aef-4e75-a127-7e82ae15a003" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.405s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.061545] env[61867]: DEBUG nova.scheduler.client.report [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 881.097484] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Volume attach. Driver type: vmdk {{(pid=61867) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 881.097764] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274369', 'volume_id': 'b788827a-5734-4c3e-a38d-82845c30c290', 'name': 'volume-b788827a-5734-4c3e-a38d-82845c30c290', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6dda389d-a8c5-4e0e-87a5-4065e24c034e', 'attached_at': '', 'detached_at': '', 'volume_id': 'b788827a-5734-4c3e-a38d-82845c30c290', 'serial': 'b788827a-5734-4c3e-a38d-82845c30c290'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 881.098665] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-586fc5bc-1f2b-474b-bb4e-b9de5929f027 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.116889] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75afaef0-805d-475a-afe6-dff114b403ca {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.148685] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] volume-b788827a-5734-4c3e-a38d-82845c30c290/volume-b788827a-5734-4c3e-a38d-82845c30c290.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.149507] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b659f4a-9b30-4628-98ea-864efbc1c0ab {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.173247] env[61867]: DEBUG oslo_vmware.api [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 881.173247] env[61867]: value = "task-1276755" [ 881.173247] env[61867]: _type = "Task" [ 881.173247] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.187738] env[61867]: DEBUG oslo_vmware.api [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276755, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.236734] env[61867]: DEBUG nova.compute.manager [req-3067556c-d975-46d3-964e-ff04f1f878cf req-cedcd6d7-b901-4c85-bf59-cae8e009a634 service nova] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Received event network-vif-plugged-41bb6052-bacd-4d6b-abd6-b5186a2418e7 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.237120] env[61867]: DEBUG oslo_concurrency.lockutils [req-3067556c-d975-46d3-964e-ff04f1f878cf req-cedcd6d7-b901-4c85-bf59-cae8e009a634 service nova] Acquiring lock "b8e561df-069e-4873-91c3-1bfe5f27de2d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.237445] env[61867]: DEBUG oslo_concurrency.lockutils [req-3067556c-d975-46d3-964e-ff04f1f878cf req-cedcd6d7-b901-4c85-bf59-cae8e009a634 service nova] Lock "b8e561df-069e-4873-91c3-1bfe5f27de2d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.237596] env[61867]: DEBUG oslo_concurrency.lockutils [req-3067556c-d975-46d3-964e-ff04f1f878cf req-cedcd6d7-b901-4c85-bf59-cae8e009a634 service nova] Lock "b8e561df-069e-4873-91c3-1bfe5f27de2d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.237777] env[61867]: DEBUG nova.compute.manager [req-3067556c-d975-46d3-964e-ff04f1f878cf req-cedcd6d7-b901-4c85-bf59-cae8e009a634 service nova] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] No waiting events found dispatching network-vif-plugged-41bb6052-bacd-4d6b-abd6-b5186a2418e7 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 881.237950] env[61867]: WARNING nova.compute.manager [req-3067556c-d975-46d3-964e-ff04f1f878cf req-cedcd6d7-b901-4c85-bf59-cae8e009a634 service nova] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Received unexpected event network-vif-plugged-41bb6052-bacd-4d6b-abd6-b5186a2418e7 for instance with vm_state building and task_state spawning. [ 881.331419] env[61867]: DEBUG nova.network.neutron [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Successfully updated port: 41bb6052-bacd-4d6b-abd6-b5186a2418e7 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 881.347191] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276754, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.360283] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276753, 'name': CreateSnapshot_Task, 'duration_secs': 0.931232} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.360581] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Created Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 881.361464] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c2977e3-a48c-42a2-b8f0-c03b5fc3790b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.567378] env[61867]: DEBUG oslo_concurrency.lockutils [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.788s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.570034] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.912s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.594047] env[61867]: INFO nova.scheduler.client.report [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleted allocations for instance 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe [ 881.687129] env[61867]: DEBUG oslo_vmware.api [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276755, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.710830] env[61867]: DEBUG nova.compute.manager [req-661cd278-f26e-483c-8188-f69c93242e5e req-b040c2bc-e1e1-4959-91db-3b6da44090ce service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Received event network-changed-4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.711168] env[61867]: DEBUG nova.compute.manager [req-661cd278-f26e-483c-8188-f69c93242e5e req-b040c2bc-e1e1-4959-91db-3b6da44090ce service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Refreshing instance network info cache due to event network-changed-4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 881.711503] env[61867]: DEBUG oslo_concurrency.lockutils [req-661cd278-f26e-483c-8188-f69c93242e5e req-b040c2bc-e1e1-4959-91db-3b6da44090ce service nova] Acquiring lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.711750] env[61867]: DEBUG oslo_concurrency.lockutils [req-661cd278-f26e-483c-8188-f69c93242e5e req-b040c2bc-e1e1-4959-91db-3b6da44090ce service nova] Acquired lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.711941] env[61867]: DEBUG nova.network.neutron [req-661cd278-f26e-483c-8188-f69c93242e5e req-b040c2bc-e1e1-4959-91db-3b6da44090ce service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Refreshing network info cache for port 4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 881.838665] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "refresh_cache-b8e561df-069e-4873-91c3-1bfe5f27de2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.838864] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquired lock "refresh_cache-b8e561df-069e-4873-91c3-1bfe5f27de2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.839248] env[61867]: DEBUG nova.network.neutron [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 881.849017] env[61867]: DEBUG oslo_vmware.api [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276754, 'name': PowerOnVM_Task, 'duration_secs': 0.976971} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.850721] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 881.850956] env[61867]: DEBUG nova.compute.manager [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 881.852839] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9e31846-0bed-4526-99c2-9b3e89946261 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.886684] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Creating linked-clone VM from snapshot {{(pid=61867) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 881.887157] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-53be62e6-aacc-47b9-a3f8-cd075c0d96f3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.896571] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 881.896571] env[61867]: value = "task-1276756" [ 881.896571] env[61867]: _type = "Task" [ 881.896571] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.906059] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276756, 'name': CloneVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.102818] env[61867]: DEBUG oslo_concurrency.lockutils [None req-185ade98-08b7-4b43-a374-01b371a3875c tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "71cd9036-0f99-4e30-aad6-ceb4f15d4ffe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.853s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.185756] env[61867]: DEBUG oslo_vmware.api [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276755, 'name': ReconfigVM_Task, 'duration_secs': 0.721629} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.186122] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Reconfigured VM instance instance-00000045 to attach disk [datastore1] volume-b788827a-5734-4c3e-a38d-82845c30c290/volume-b788827a-5734-4c3e-a38d-82845c30c290.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.191633] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02cf96de-3d83-43a0-8752-72e9db60fe88 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.210391] env[61867]: DEBUG oslo_vmware.api [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 882.210391] env[61867]: value = "task-1276757" [ 882.210391] env[61867]: _type = "Task" [ 882.210391] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.223262] env[61867]: DEBUG oslo_vmware.api [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276757, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.378418] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.397144] env[61867]: DEBUG nova.network.neutron [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 882.412291] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276756, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.567925] env[61867]: DEBUG nova.network.neutron [req-661cd278-f26e-483c-8188-f69c93242e5e req-b040c2bc-e1e1-4959-91db-3b6da44090ce service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Updated VIF entry in instance network info cache for port 4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 882.568447] env[61867]: DEBUG nova.network.neutron [req-661cd278-f26e-483c-8188-f69c93242e5e req-b040c2bc-e1e1-4959-91db-3b6da44090ce service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Updating instance_info_cache with network_info: [{"id": "4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1", "address": "fa:16:3e:ce:d6:b5", "network": {"id": "ab6ec5b9-8193-43c9-a3ce-d63f31e5a001", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1573384141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35aa85105b394369b17ef4ce51a49066", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f63bd0e-c1", "ovs_interfaceid": "4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.601466] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance abb41c0c-6d0d-4147-a4af-554ab7d9e921 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.601643] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 6e41989e-b8fa-4009-af1e-1ce859b329a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.601771] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 8a83f4f2-58eb-473e-9b1e-32ce633554f9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.601902] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 6dda389d-a8c5-4e0e-87a5-4065e24c034e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.602035] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance ca2f6d18-f773-4875-83de-2f2be912c2f8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.602231] env[61867]: WARNING nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 81e70c36-04ea-450c-9383-53ef069d1c46 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 882.602359] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 6d2dab88-4165-4952-8019-2eaf3b863115 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.602480] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 2aa08603-d87f-4734-bdfe-fdd610d54e1f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.602593] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 07fdd75c-34ab-45e5-a98b-aced2caa6cb5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.602707] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 7479bf91-5aef-4e75-a127-7e82ae15a003 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.602826] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance b8e561df-069e-4873-91c3-1bfe5f27de2d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.623076] env[61867]: DEBUG nova.network.neutron [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Updating instance_info_cache with network_info: [{"id": "41bb6052-bacd-4d6b-abd6-b5186a2418e7", "address": "fa:16:3e:5e:18:38", "network": {"id": "a645da41-1400-470e-acf2-68cec86432fd", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1387586368-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a29c5e9cfdaa4cc88ed300100a308ffa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41bb6052-ba", "ovs_interfaceid": "41bb6052-bacd-4d6b-abd6-b5186a2418e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.722089] env[61867]: DEBUG oslo_vmware.api [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276757, 'name': ReconfigVM_Task, 'duration_secs': 0.185949} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.722089] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274369', 'volume_id': 'b788827a-5734-4c3e-a38d-82845c30c290', 'name': 'volume-b788827a-5734-4c3e-a38d-82845c30c290', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6dda389d-a8c5-4e0e-87a5-4065e24c034e', 'attached_at': '', 'detached_at': '', 'volume_id': 'b788827a-5734-4c3e-a38d-82845c30c290', 'serial': 'b788827a-5734-4c3e-a38d-82845c30c290'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 882.909482] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276756, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.071440] env[61867]: DEBUG oslo_concurrency.lockutils [req-661cd278-f26e-483c-8188-f69c93242e5e req-b040c2bc-e1e1-4959-91db-3b6da44090ce service nova] Releasing lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.105713] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 1df8427c-e75d-4b60-a92a-b5ba76b67081 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 883.126010] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Releasing lock "refresh_cache-b8e561df-069e-4873-91c3-1bfe5f27de2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.126352] env[61867]: DEBUG nova.compute.manager [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Instance network_info: |[{"id": "41bb6052-bacd-4d6b-abd6-b5186a2418e7", "address": "fa:16:3e:5e:18:38", "network": {"id": "a645da41-1400-470e-acf2-68cec86432fd", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1387586368-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a29c5e9cfdaa4cc88ed300100a308ffa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41bb6052-ba", "ovs_interfaceid": "41bb6052-bacd-4d6b-abd6-b5186a2418e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 883.127030] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:18:38', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d377d75-3add-4a15-8691-74b2eb010924', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '41bb6052-bacd-4d6b-abd6-b5186a2418e7', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 883.135708] env[61867]: DEBUG oslo.service.loopingcall [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.135866] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 883.136196] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-af9e1ff1-0b9e-4e03-a1db-06487415c9c2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.158029] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 883.158029] env[61867]: value = "task-1276758" [ 883.158029] env[61867]: _type = "Task" [ 883.158029] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.170701] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276758, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.325054] env[61867]: DEBUG nova.compute.manager [req-8e83a53d-13eb-4e64-a456-143bb6692fff req-e7771095-8f1a-4846-b8c1-7e82adcabf92 service nova] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Received event network-changed-41bb6052-bacd-4d6b-abd6-b5186a2418e7 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.326398] env[61867]: DEBUG nova.compute.manager [req-8e83a53d-13eb-4e64-a456-143bb6692fff req-e7771095-8f1a-4846-b8c1-7e82adcabf92 service nova] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Refreshing instance network info cache due to event network-changed-41bb6052-bacd-4d6b-abd6-b5186a2418e7. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 883.326398] env[61867]: DEBUG oslo_concurrency.lockutils [req-8e83a53d-13eb-4e64-a456-143bb6692fff req-e7771095-8f1a-4846-b8c1-7e82adcabf92 service nova] Acquiring lock "refresh_cache-b8e561df-069e-4873-91c3-1bfe5f27de2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.326398] env[61867]: DEBUG oslo_concurrency.lockutils [req-8e83a53d-13eb-4e64-a456-143bb6692fff req-e7771095-8f1a-4846-b8c1-7e82adcabf92 service nova] Acquired lock "refresh_cache-b8e561df-069e-4873-91c3-1bfe5f27de2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.326398] env[61867]: DEBUG nova.network.neutron [req-8e83a53d-13eb-4e64-a456-143bb6692fff req-e7771095-8f1a-4846-b8c1-7e82adcabf92 service nova] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Refreshing network info cache for port 41bb6052-bacd-4d6b-abd6-b5186a2418e7 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 883.410358] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276756, 'name': CloneVM_Task, 'duration_secs': 1.47599} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.410722] env[61867]: INFO nova.virt.vmwareapi.vmops [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Created linked-clone VM from snapshot [ 883.411430] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce04cf46-3135-4a2c-a888-509296b21419 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.419820] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Uploading image 928c7641-2c77-4753-814a-9eb159d83adf {{(pid=61867) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 883.441924] env[61867]: DEBUG oslo_vmware.rw_handles [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 883.441924] env[61867]: value = "vm-274371" [ 883.441924] env[61867]: _type = "VirtualMachine" [ 883.441924] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 883.442271] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-64ecc04a-59f2-4517-96cc-67de24ecbbed {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.450513] env[61867]: DEBUG oslo_vmware.rw_handles [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lease: (returnval){ [ 883.450513] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52464637-358b-fbc7-6ff1-9f7c31b2151d" [ 883.450513] env[61867]: _type = "HttpNfcLease" [ 883.450513] env[61867]: } obtained for exporting VM: (result){ [ 883.450513] env[61867]: value = "vm-274371" [ 883.450513] env[61867]: _type = "VirtualMachine" [ 883.450513] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 883.450801] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the lease: (returnval){ [ 883.450801] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52464637-358b-fbc7-6ff1-9f7c31b2151d" [ 883.450801] env[61867]: _type = "HttpNfcLease" [ 883.450801] env[61867]: } to be ready. {{(pid=61867) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 883.458071] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 883.458071] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52464637-358b-fbc7-6ff1-9f7c31b2151d" [ 883.458071] env[61867]: _type = "HttpNfcLease" [ 883.458071] env[61867]: } is initializing. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 883.609462] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance adc2732d-2a10-40ce-bb90-ed0762a36614 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 883.609662] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 883.609787] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2432MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 883.619502] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.619734] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.672865] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276758, 'name': CreateVM_Task} progress is 25%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.764113] env[61867]: DEBUG nova.objects.instance [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lazy-loading 'flavor' on Instance uuid 6dda389d-a8c5-4e0e-87a5-4065e24c034e {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.801734] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a11acda-65a5-4466-a5d4-96675c10ee43 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.810029] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23428b94-2956-447b-8599-eb4e25f3e22d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.841696] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-135161be-bca4-4eff-919a-452d1d62be2e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.850584] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d932b2ac-4195-431e-9bc5-8f79b310b4a7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.865463] env[61867]: DEBUG nova.compute.provider_tree [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.885731] env[61867]: DEBUG oslo_concurrency.lockutils [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "07fdd75c-34ab-45e5-a98b-aced2caa6cb5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.885988] env[61867]: DEBUG oslo_concurrency.lockutils [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "07fdd75c-34ab-45e5-a98b-aced2caa6cb5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.886333] env[61867]: DEBUG oslo_concurrency.lockutils [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "07fdd75c-34ab-45e5-a98b-aced2caa6cb5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.886424] env[61867]: DEBUG oslo_concurrency.lockutils [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "07fdd75c-34ab-45e5-a98b-aced2caa6cb5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.886578] env[61867]: DEBUG oslo_concurrency.lockutils [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "07fdd75c-34ab-45e5-a98b-aced2caa6cb5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.888535] env[61867]: INFO nova.compute.manager [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Terminating instance [ 883.890272] env[61867]: DEBUG nova.compute.manager [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 883.890472] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 883.891283] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50dc28c-f974-40bb-9bda-6267c34caaa2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.899126] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 883.899365] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-16bf27f9-11fa-45b1-ad6b-b66160f50d6f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.906362] env[61867]: DEBUG oslo_vmware.api [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 883.906362] env[61867]: value = "task-1276760" [ 883.906362] env[61867]: _type = "Task" [ 883.906362] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.918433] env[61867]: DEBUG oslo_vmware.api [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276760, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.959957] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 883.959957] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52464637-358b-fbc7-6ff1-9f7c31b2151d" [ 883.959957] env[61867]: _type = "HttpNfcLease" [ 883.959957] env[61867]: } is ready. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 883.960412] env[61867]: DEBUG oslo_vmware.rw_handles [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 883.960412] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52464637-358b-fbc7-6ff1-9f7c31b2151d" [ 883.960412] env[61867]: _type = "HttpNfcLease" [ 883.960412] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 883.961182] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db257be-1508-4751-80e5-c8f18cf2bacb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.969342] env[61867]: DEBUG oslo_vmware.rw_handles [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c96c07-5cfb-cbcb-19e0-6f10d3b78187/disk-0.vmdk from lease info. {{(pid=61867) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 883.969537] env[61867]: DEBUG oslo_vmware.rw_handles [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c96c07-5cfb-cbcb-19e0-6f10d3b78187/disk-0.vmdk for reading. {{(pid=61867) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 884.061427] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-acb314ec-0b10-4e0a-9c70-c8f365fc3e77 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.120719] env[61867]: DEBUG nova.network.neutron [req-8e83a53d-13eb-4e64-a456-143bb6692fff req-e7771095-8f1a-4846-b8c1-7e82adcabf92 service nova] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Updated VIF entry in instance network info cache for port 41bb6052-bacd-4d6b-abd6-b5186a2418e7. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 884.121189] env[61867]: DEBUG nova.network.neutron [req-8e83a53d-13eb-4e64-a456-143bb6692fff req-e7771095-8f1a-4846-b8c1-7e82adcabf92 service nova] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Updating instance_info_cache with network_info: [{"id": "41bb6052-bacd-4d6b-abd6-b5186a2418e7", "address": "fa:16:3e:5e:18:38", "network": {"id": "a645da41-1400-470e-acf2-68cec86432fd", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1387586368-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a29c5e9cfdaa4cc88ed300100a308ffa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41bb6052-ba", "ovs_interfaceid": "41bb6052-bacd-4d6b-abd6-b5186a2418e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.122542] env[61867]: DEBUG nova.compute.manager [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 884.170852] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276758, 'name': CreateVM_Task, 'duration_secs': 0.678514} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.171046] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 884.171807] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.172030] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.172439] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 884.172696] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91a31a93-e40a-4601-bb80-684c19a93cba {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.178656] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 884.178656] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52dbe948-638e-b36c-9d67-9e9507cf1137" [ 884.178656] env[61867]: _type = "Task" [ 884.178656] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.187589] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52dbe948-638e-b36c-9d67-9e9507cf1137, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.272761] env[61867]: DEBUG oslo_concurrency.lockutils [None req-210a140d-2f9a-4b4b-b211-2cf710705142 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.830s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.367295] env[61867]: DEBUG nova.scheduler.client.report [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.375907] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.376172] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.420593] env[61867]: DEBUG oslo_vmware.api [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276760, 'name': PowerOffVM_Task, 'duration_secs': 0.196084} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.421184] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 884.421417] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 884.421786] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-189e0e6d-28f8-4229-9b78-3242e7178474 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.518446] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 884.518771] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 884.519061] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleting the datastore file [datastore1] 07fdd75c-34ab-45e5-a98b-aced2caa6cb5 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 884.519367] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4f06dfaf-c8b7-4a57-ba9e-f8e77cd099d6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.527954] env[61867]: DEBUG oslo_vmware.api [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 884.527954] env[61867]: value = "task-1276762" [ 884.527954] env[61867]: _type = "Task" [ 884.527954] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.538033] env[61867]: DEBUG oslo_vmware.api [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276762, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.626776] env[61867]: DEBUG oslo_concurrency.lockutils [req-8e83a53d-13eb-4e64-a456-143bb6692fff req-e7771095-8f1a-4846-b8c1-7e82adcabf92 service nova] Releasing lock "refresh_cache-b8e561df-069e-4873-91c3-1bfe5f27de2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.653313] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.689951] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52dbe948-638e-b36c-9d67-9e9507cf1137, 'name': SearchDatastore_Task, 'duration_secs': 0.026823} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.690542] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.690831] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 884.691178] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.691428] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.691681] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 884.692537] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-294c39ad-6773-4ce5-8a3f-e1161f694ce4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.711951] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 884.712342] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 884.713797] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21309698-9284-451d-8b74-cf982bf5ced6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.720624] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 884.720624] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d6e47f-3c00-4e5c-a5a8-d27f5ccaede7" [ 884.720624] env[61867]: _type = "Task" [ 884.720624] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.731975] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d6e47f-3c00-4e5c-a5a8-d27f5ccaede7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.872743] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61867) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 884.872974] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.303s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.873357] env[61867]: DEBUG oslo_concurrency.lockutils [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.380s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.875081] env[61867]: INFO nova.compute.claims [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 884.878444] env[61867]: DEBUG nova.compute.manager [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 885.040778] env[61867]: DEBUG oslo_vmware.api [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276762, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.296396} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.041365] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 885.042013] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 885.042013] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 885.042108] env[61867]: INFO nova.compute.manager [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Took 1.15 seconds to destroy the instance on the hypervisor. [ 885.042475] env[61867]: DEBUG oslo.service.loopingcall [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.042544] env[61867]: DEBUG nova.compute.manager [-] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 885.042635] env[61867]: DEBUG nova.network.neutron [-] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 885.149090] env[61867]: DEBUG oslo_concurrency.lockutils [None req-31cbbb5d-821c-445e-9460-259a36980aa2 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "interface-ca2f6d18-f773-4875-83de-2f2be912c2f8-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.149594] env[61867]: DEBUG oslo_concurrency.lockutils [None req-31cbbb5d-821c-445e-9460-259a36980aa2 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "interface-ca2f6d18-f773-4875-83de-2f2be912c2f8-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.150054] env[61867]: DEBUG nova.objects.instance [None req-31cbbb5d-821c-445e-9460-259a36980aa2 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lazy-loading 'flavor' on Instance uuid ca2f6d18-f773-4875-83de-2f2be912c2f8 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.231188] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d6e47f-3c00-4e5c-a5a8-d27f5ccaede7, 'name': SearchDatastore_Task, 'duration_secs': 0.010343} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.232129] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1782a7ab-5229-40ff-ad44-d6a03444b8d2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.239733] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 885.239733] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5294c67d-bd90-a3fa-da29-0b8b1faf0dc4" [ 885.239733] env[61867]: _type = "Task" [ 885.239733] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.248047] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5294c67d-bd90-a3fa-da29-0b8b1faf0dc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.351070] env[61867]: DEBUG nova.compute.manager [req-9ebde61f-409b-4f03-82a8-30283cade2ae req-0fa752ed-806e-4f2f-a24e-bc6279308432 service nova] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Received event network-vif-deleted-2799f597-62eb-4c48-929e-69844f6e0dea {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.351423] env[61867]: INFO nova.compute.manager [req-9ebde61f-409b-4f03-82a8-30283cade2ae req-0fa752ed-806e-4f2f-a24e-bc6279308432 service nova] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Neutron deleted interface 2799f597-62eb-4c48-929e-69844f6e0dea; detaching it from the instance and deleting it from the info cache [ 885.351672] env[61867]: DEBUG nova.network.neutron [req-9ebde61f-409b-4f03-82a8-30283cade2ae req-0fa752ed-806e-4f2f-a24e-bc6279308432 service nova] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.403330] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.654436] env[61867]: DEBUG nova.objects.instance [None req-31cbbb5d-821c-445e-9460-259a36980aa2 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lazy-loading 'pci_requests' on Instance uuid ca2f6d18-f773-4875-83de-2f2be912c2f8 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.667701] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.668019] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.751844] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5294c67d-bd90-a3fa-da29-0b8b1faf0dc4, 'name': SearchDatastore_Task, 'duration_secs': 0.015562} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.752745] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.752982] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] b8e561df-069e-4873-91c3-1bfe5f27de2d/b8e561df-069e-4873-91c3-1bfe5f27de2d.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 885.753300] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6eb723e5-7e12-4a0c-97bb-6da65476ba79 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.762413] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 885.762413] env[61867]: value = "task-1276763" [ 885.762413] env[61867]: _type = "Task" [ 885.762413] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.771858] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276763, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.805448] env[61867]: DEBUG nova.network.neutron [-] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.856407] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c2ed0c85-4384-49f7-ba26-b16984cd11d9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.869323] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9376393b-fc11-46d5-9bf2-d312df49391d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.908222] env[61867]: DEBUG nova.compute.manager [req-9ebde61f-409b-4f03-82a8-30283cade2ae req-0fa752ed-806e-4f2f-a24e-bc6279308432 service nova] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Detach interface failed, port_id=2799f597-62eb-4c48-929e-69844f6e0dea, reason: Instance 07fdd75c-34ab-45e5-a98b-aced2caa6cb5 could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 886.106057] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2e42e0-20d2-4a70-a804-86260ce11203 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.115861] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31eb0f86-c755-43ac-811a-8c46b9cd9cde {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.149666] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-671590fa-6ee0-4554-90dd-cf47febef8f0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.158063] env[61867]: DEBUG nova.objects.base [None req-31cbbb5d-821c-445e-9460-259a36980aa2 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61867) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 886.158063] env[61867]: DEBUG nova.network.neutron [None req-31cbbb5d-821c-445e-9460-259a36980aa2 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 886.164442] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3901b49-f96c-4ce8-af48-0655a76f1e3e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.171479] env[61867]: DEBUG nova.compute.utils [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 886.191498] env[61867]: DEBUG nova.compute.provider_tree [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.260818] env[61867]: DEBUG oslo_concurrency.lockutils [None req-31cbbb5d-821c-445e-9460-259a36980aa2 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "interface-ca2f6d18-f773-4875-83de-2f2be912c2f8-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.111s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.272990] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276763, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.308530] env[61867]: INFO nova.compute.manager [-] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Took 1.27 seconds to deallocate network for instance. [ 886.675175] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.696086] env[61867]: DEBUG nova.scheduler.client.report [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.776856] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276763, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555474} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.777270] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] b8e561df-069e-4873-91c3-1bfe5f27de2d/b8e561df-069e-4873-91c3-1bfe5f27de2d.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 886.777549] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 886.777877] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3d673441-6022-4e31-922d-2bb656f1a436 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.787686] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 886.787686] env[61867]: value = "task-1276764" [ 886.787686] env[61867]: _type = "Task" [ 886.787686] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.798273] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276764, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.816146] env[61867]: DEBUG oslo_concurrency.lockutils [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.200997] env[61867]: DEBUG oslo_concurrency.lockutils [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.328s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.201660] env[61867]: DEBUG nova.compute.manager [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 887.204361] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.115s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.204542] env[61867]: DEBUG nova.objects.instance [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lazy-loading 'pci_requests' on Instance uuid 1df8427c-e75d-4b60-a92a-b5ba76b67081 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.298600] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276764, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082614} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.298919] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 887.299752] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab9e62fb-4040-462b-82ea-41798cb19ec4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.323216] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] b8e561df-069e-4873-91c3-1bfe5f27de2d/b8e561df-069e-4873-91c3-1bfe5f27de2d.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 887.323598] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de4e8783-ec44-4f7e-b6dc-6cf9bd11c91c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.345181] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 887.345181] env[61867]: value = "task-1276765" [ 887.345181] env[61867]: _type = "Task" [ 887.345181] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.355569] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276765, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.707321] env[61867]: DEBUG nova.compute.utils [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 887.710842] env[61867]: DEBUG nova.objects.instance [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lazy-loading 'numa_topology' on Instance uuid 1df8427c-e75d-4b60-a92a-b5ba76b67081 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.712014] env[61867]: DEBUG nova.compute.manager [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 887.712205] env[61867]: DEBUG nova.network.neutron [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 887.740633] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.740902] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.741274] env[61867]: INFO nova.compute.manager [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Attaching volume 14e9fde5-2e48-4463-8391-cf95a9fa600a to /dev/sdc [ 887.753212] env[61867]: DEBUG nova.policy [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'adcd9eb75ecc4eccb335ebfec207a900', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7316e4f263a9432ab2f9f91484d62f58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 887.773199] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e94c1e9-8d7b-4ef7-9512-64c43026dcee {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.781314] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e7ece31-3bac-4aac-b322-dd0dc718ea5a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.797260] env[61867]: DEBUG nova.virt.block_device [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Updating existing volume attachment record: add8c798-2b0c-451e-94d8-d4adf49623e9 {{(pid=61867) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 887.856548] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276765, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.014032] env[61867]: DEBUG nova.network.neutron [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Successfully created port: a101c251-3c8e-4c0f-8528-aba0e8002f42 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 888.188621] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "interface-ca2f6d18-f773-4875-83de-2f2be912c2f8-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.188943] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "interface-ca2f6d18-f773-4875-83de-2f2be912c2f8-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.189385] env[61867]: DEBUG nova.objects.instance [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lazy-loading 'flavor' on Instance uuid ca2f6d18-f773-4875-83de-2f2be912c2f8 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.212667] env[61867]: DEBUG nova.compute.manager [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 888.218557] env[61867]: INFO nova.compute.claims [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 888.358053] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276765, 'name': ReconfigVM_Task, 'duration_secs': 0.677153} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.358053] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Reconfigured VM instance instance-00000051 to attach disk [datastore2] b8e561df-069e-4873-91c3-1bfe5f27de2d/b8e561df-069e-4873-91c3-1bfe5f27de2d.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 888.358570] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-28e830ed-0663-4611-8667-62505f077771 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.367621] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 888.367621] env[61867]: value = "task-1276767" [ 888.367621] env[61867]: _type = "Task" [ 888.367621] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.377321] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276767, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.765937] env[61867]: DEBUG nova.objects.instance [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lazy-loading 'pci_requests' on Instance uuid ca2f6d18-f773-4875-83de-2f2be912c2f8 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.879053] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276767, 'name': Rename_Task, 'duration_secs': 0.164376} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.879259] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 888.879533] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-145c2492-cf9d-404f-b57b-909a0d1c9cdb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.886692] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 888.886692] env[61867]: value = "task-1276768" [ 888.886692] env[61867]: _type = "Task" [ 888.886692] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.895177] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276768, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.227038] env[61867]: DEBUG nova.compute.manager [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 889.249993] env[61867]: DEBUG nova.virt.hardware [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 889.250294] env[61867]: DEBUG nova.virt.hardware [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 889.250466] env[61867]: DEBUG nova.virt.hardware [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 889.250646] env[61867]: DEBUG nova.virt.hardware [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 889.250796] env[61867]: DEBUG nova.virt.hardware [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 889.250948] env[61867]: DEBUG nova.virt.hardware [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 889.251215] env[61867]: DEBUG nova.virt.hardware [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 889.251398] env[61867]: DEBUG nova.virt.hardware [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 889.251574] env[61867]: DEBUG nova.virt.hardware [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 889.251742] env[61867]: DEBUG nova.virt.hardware [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 889.251915] env[61867]: DEBUG nova.virt.hardware [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 889.252858] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec70ae2e-8f24-4d06-bc82-95559452bb34 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.264350] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682ac7df-2dea-46c2-85c5-925fc2345e1c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.268398] env[61867]: DEBUG nova.objects.base [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61867) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 889.268569] env[61867]: DEBUG nova.network.neutron [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 889.344783] env[61867]: DEBUG nova.policy [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ea634246436422a9407f82692b69ede', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '86ced20eaf4740e298dc6f8ca5550c09', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 889.399492] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276768, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.455627] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff20b2f-8a08-4443-bb82-9d7650c21b8b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.466782] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7330d893-0cc2-40f8-af94-d29686939ee9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.501700] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0caf3a8-b16a-4631-b70f-af181e27ba56 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.509856] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c69ada1a-3eb9-4bd5-9a06-3f2c33deb291 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.524197] env[61867]: DEBUG nova.compute.provider_tree [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 889.548440] env[61867]: DEBUG nova.compute.manager [req-a1651803-7f6d-49fb-86d3-14d454047019 req-2080c7a8-4c9c-4889-82db-5ba3bb3ad7a2 service nova] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Received event network-vif-plugged-a101c251-3c8e-4c0f-8528-aba0e8002f42 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.548667] env[61867]: DEBUG oslo_concurrency.lockutils [req-a1651803-7f6d-49fb-86d3-14d454047019 req-2080c7a8-4c9c-4889-82db-5ba3bb3ad7a2 service nova] Acquiring lock "adc2732d-2a10-40ce-bb90-ed0762a36614-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.548886] env[61867]: DEBUG oslo_concurrency.lockutils [req-a1651803-7f6d-49fb-86d3-14d454047019 req-2080c7a8-4c9c-4889-82db-5ba3bb3ad7a2 service nova] Lock "adc2732d-2a10-40ce-bb90-ed0762a36614-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.549074] env[61867]: DEBUG oslo_concurrency.lockutils [req-a1651803-7f6d-49fb-86d3-14d454047019 req-2080c7a8-4c9c-4889-82db-5ba3bb3ad7a2 service nova] Lock "adc2732d-2a10-40ce-bb90-ed0762a36614-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.549388] env[61867]: DEBUG nova.compute.manager [req-a1651803-7f6d-49fb-86d3-14d454047019 req-2080c7a8-4c9c-4889-82db-5ba3bb3ad7a2 service nova] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] No waiting events found dispatching network-vif-plugged-a101c251-3c8e-4c0f-8528-aba0e8002f42 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 889.549506] env[61867]: WARNING nova.compute.manager [req-a1651803-7f6d-49fb-86d3-14d454047019 req-2080c7a8-4c9c-4889-82db-5ba3bb3ad7a2 service nova] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Received unexpected event network-vif-plugged-a101c251-3c8e-4c0f-8528-aba0e8002f42 for instance with vm_state building and task_state spawning. [ 889.770229] env[61867]: DEBUG nova.network.neutron [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Successfully updated port: a101c251-3c8e-4c0f-8528-aba0e8002f42 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 889.837494] env[61867]: DEBUG nova.network.neutron [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Successfully created port: 65189f98-8c73-47e1-94cd-60c94c3a65d7 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 889.899567] env[61867]: DEBUG oslo_vmware.api [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276768, 'name': PowerOnVM_Task, 'duration_secs': 0.569341} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.899854] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 889.900083] env[61867]: INFO nova.compute.manager [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Took 8.97 seconds to spawn the instance on the hypervisor. [ 889.900275] env[61867]: DEBUG nova.compute.manager [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 889.901118] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-122ccd8c-76d2-4a64-898d-28dc6fe6398b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.027670] env[61867]: DEBUG nova.scheduler.client.report [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.272827] env[61867]: DEBUG oslo_concurrency.lockutils [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "refresh_cache-adc2732d-2a10-40ce-bb90-ed0762a36614" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.273204] env[61867]: DEBUG oslo_concurrency.lockutils [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "refresh_cache-adc2732d-2a10-40ce-bb90-ed0762a36614" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.273302] env[61867]: DEBUG nova.network.neutron [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 890.420956] env[61867]: INFO nova.compute.manager [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Took 16.66 seconds to build instance. [ 890.533251] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.329s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.536038] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.528s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.536038] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.537809] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 8.159s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.538044] env[61867]: DEBUG nova.objects.instance [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61867) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 890.562253] env[61867]: INFO nova.scheduler.client.report [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Deleted allocations for instance 81e70c36-04ea-450c-9383-53ef069d1c46 [ 890.571029] env[61867]: INFO nova.network.neutron [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Updating port 943c04a5-7eba-46db-ab4f-93f7f84be14e with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 890.805692] env[61867]: DEBUG nova.network.neutron [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 890.923655] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c1edc850-4a9e-4079-863c-2e254b7f24a5 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "b8e561df-069e-4873-91c3-1bfe5f27de2d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.176s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.003889] env[61867]: DEBUG nova.network.neutron [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Updating instance_info_cache with network_info: [{"id": "a101c251-3c8e-4c0f-8528-aba0e8002f42", "address": "fa:16:3e:6f:bc:12", "network": {"id": "14cee476-296a-4093-99ba-c2a5ef63f7cf", "bridge": "br-int", "label": "tempest-ServersTestJSON-243183479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7316e4f263a9432ab2f9f91484d62f58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa101c251-3c", "ovs_interfaceid": "a101c251-3c8e-4c0f-8528-aba0e8002f42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.070576] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1458da1b-a719-4d39-872d-175f82e25798 tempest-MigrationsAdminTest-2097416287 tempest-MigrationsAdminTest-2097416287-project-member] Lock "81e70c36-04ea-450c-9383-53ef069d1c46" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.010s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.358468] env[61867]: DEBUG nova.network.neutron [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Successfully updated port: 65189f98-8c73-47e1-94cd-60c94c3a65d7 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 891.506885] env[61867]: DEBUG oslo_concurrency.lockutils [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "refresh_cache-adc2732d-2a10-40ce-bb90-ed0762a36614" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.506974] env[61867]: DEBUG nova.compute.manager [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Instance network_info: |[{"id": "a101c251-3c8e-4c0f-8528-aba0e8002f42", "address": "fa:16:3e:6f:bc:12", "network": {"id": "14cee476-296a-4093-99ba-c2a5ef63f7cf", "bridge": "br-int", "label": "tempest-ServersTestJSON-243183479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7316e4f263a9432ab2f9f91484d62f58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa101c251-3c", "ovs_interfaceid": "a101c251-3c8e-4c0f-8528-aba0e8002f42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 891.507438] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:bc:12', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '085fb0ff-9285-4f1d-a008-a14da4844357', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a101c251-3c8e-4c0f-8528-aba0e8002f42', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 891.515864] env[61867]: DEBUG oslo.service.loopingcall [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 891.516460] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 891.516713] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-99c8af26-9878-4819-8a0d-31d77c07d58b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.540466] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 891.540466] env[61867]: value = "task-1276770" [ 891.540466] env[61867]: _type = "Task" [ 891.540466] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.546558] env[61867]: DEBUG oslo_concurrency.lockutils [None req-ed9aa042-e0a0-497e-a89a-78664ae96055 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.551850] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.899s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.554529] env[61867]: INFO nova.compute.claims [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.556956] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276770, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.598104] env[61867]: DEBUG nova.compute.manager [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Received event network-changed-a101c251-3c8e-4c0f-8528-aba0e8002f42 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 891.598357] env[61867]: DEBUG nova.compute.manager [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Refreshing instance network info cache due to event network-changed-a101c251-3c8e-4c0f-8528-aba0e8002f42. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 891.598979] env[61867]: DEBUG oslo_concurrency.lockutils [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] Acquiring lock "refresh_cache-adc2732d-2a10-40ce-bb90-ed0762a36614" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.599165] env[61867]: DEBUG oslo_concurrency.lockutils [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] Acquired lock "refresh_cache-adc2732d-2a10-40ce-bb90-ed0762a36614" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.599388] env[61867]: DEBUG nova.network.neutron [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Refreshing network info cache for port a101c251-3c8e-4c0f-8528-aba0e8002f42 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 891.861662] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.862122] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.862122] env[61867]: DEBUG nova.network.neutron [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 892.053493] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276770, 'name': CreateVM_Task, 'duration_secs': 0.368008} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.054114] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 892.055103] env[61867]: DEBUG oslo_concurrency.lockutils [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.055103] env[61867]: DEBUG oslo_concurrency.lockutils [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.057027] env[61867]: DEBUG oslo_concurrency.lockutils [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 892.057027] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebb091de-0bd4-4e99-8332-2f252b5d7ac5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.064496] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 892.064496] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5210fc9c-bd8b-e311-ca24-a672df24b9e7" [ 892.064496] env[61867]: _type = "Task" [ 892.064496] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.078023] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5210fc9c-bd8b-e311-ca24-a672df24b9e7, 'name': SearchDatastore_Task, 'duration_secs': 0.010289} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.078023] env[61867]: DEBUG oslo_concurrency.lockutils [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.078023] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 892.078023] env[61867]: DEBUG oslo_concurrency.lockutils [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.078273] env[61867]: DEBUG oslo_concurrency.lockutils [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.078273] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 892.078273] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-242ed23e-03ae-4fac-8d40-ac9bfee7d99c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.088455] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 892.088660] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 892.089950] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c11b63cc-8e9c-43e7-9a2c-8ad6ddc0fc85 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.098467] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 892.098467] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525bcbcd-868c-86cf-a8f3-6c06c708a9e8" [ 892.098467] env[61867]: _type = "Task" [ 892.098467] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.113471] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525bcbcd-868c-86cf-a8f3-6c06c708a9e8, 'name': SearchDatastore_Task, 'duration_secs': 0.010726} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.113471] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6b4eb5d-f3db-4dd3-b60f-818100252f61 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.118747] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 892.118747] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ba2e0b-09a6-0fc5-0f80-25bfd685f454" [ 892.118747] env[61867]: _type = "Task" [ 892.118747] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.128936] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ba2e0b-09a6-0fc5-0f80-25bfd685f454, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.344719] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Volume attach. Driver type: vmdk {{(pid=61867) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 892.345072] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274373', 'volume_id': '14e9fde5-2e48-4463-8391-cf95a9fa600a', 'name': 'volume-14e9fde5-2e48-4463-8391-cf95a9fa600a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6dda389d-a8c5-4e0e-87a5-4065e24c034e', 'attached_at': '', 'detached_at': '', 'volume_id': '14e9fde5-2e48-4463-8391-cf95a9fa600a', 'serial': '14e9fde5-2e48-4463-8391-cf95a9fa600a'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 892.345975] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d55bc08-bfd2-4da3-9f1f-e1780f79377d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.370276] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d59f92e-8a48-41c0-9ab6-06a32a0a8b85 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.403758] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] volume-14e9fde5-2e48-4463-8391-cf95a9fa600a/volume-14e9fde5-2e48-4463-8391-cf95a9fa600a.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.405453] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1182205f-ef69-4037-aff8-c2e4f0c2d41f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.420670] env[61867]: WARNING nova.network.neutron [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] 7a62cd67-4e7d-4952-9a24-053e35c1c3cb already exists in list: networks containing: ['7a62cd67-4e7d-4952-9a24-053e35c1c3cb']. ignoring it [ 892.423015] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.423643] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquired lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.423833] env[61867]: DEBUG nova.network.neutron [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 892.433342] env[61867]: DEBUG oslo_vmware.api [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 892.433342] env[61867]: value = "task-1276771" [ 892.433342] env[61867]: _type = "Task" [ 892.433342] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.445719] env[61867]: DEBUG oslo_vmware.api [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276771, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.483450] env[61867]: DEBUG nova.network.neutron [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Updated VIF entry in instance network info cache for port a101c251-3c8e-4c0f-8528-aba0e8002f42. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 892.483912] env[61867]: DEBUG nova.network.neutron [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Updating instance_info_cache with network_info: [{"id": "a101c251-3c8e-4c0f-8528-aba0e8002f42", "address": "fa:16:3e:6f:bc:12", "network": {"id": "14cee476-296a-4093-99ba-c2a5ef63f7cf", "bridge": "br-int", "label": "tempest-ServersTestJSON-243183479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7316e4f263a9432ab2f9f91484d62f58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa101c251-3c", "ovs_interfaceid": "a101c251-3c8e-4c0f-8528-aba0e8002f42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.491863] env[61867]: DEBUG nova.compute.manager [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 892.492787] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61d3af13-1dfc-4f9a-8809-10adb677e5ff {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.633209] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ba2e0b-09a6-0fc5-0f80-25bfd685f454, 'name': SearchDatastore_Task, 'duration_secs': 0.009712} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.633209] env[61867]: DEBUG oslo_concurrency.lockutils [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.633489] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] adc2732d-2a10-40ce-bb90-ed0762a36614/adc2732d-2a10-40ce-bb90-ed0762a36614.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.636288] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-088a55d8-cd2f-4c58-9eb5-d5aa3f30872e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.642092] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 892.642092] env[61867]: value = "task-1276772" [ 892.642092] env[61867]: _type = "Task" [ 892.642092] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.653785] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276772, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.782012] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b836a7-c266-43d3-a2b2-e0018d0077e7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.792867] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd31851-9299-40d6-8a0d-a006962f09db {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.825095] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21e376f0-12ba-4203-9f5d-6cf1100f2c72 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.834208] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23eec29e-2cae-42b2-9e74-fd58c3197cfd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.839250] env[61867]: DEBUG nova.network.neutron [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Updating instance_info_cache with network_info: [{"id": "af89c243-c119-47c9-9eda-60c9be2b8c51", "address": "fa:16:3e:4d:1f:76", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf89c243-c1", "ovs_interfaceid": "af89c243-c119-47c9-9eda-60c9be2b8c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "65189f98-8c73-47e1-94cd-60c94c3a65d7", "address": "fa:16:3e:f2:05:68", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65189f98-8c", "ovs_interfaceid": "65189f98-8c73-47e1-94cd-60c94c3a65d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.852711] env[61867]: DEBUG nova.compute.provider_tree [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.946621] env[61867]: DEBUG oslo_vmware.api [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276771, 'name': ReconfigVM_Task, 'duration_secs': 0.488869} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.946946] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Reconfigured VM instance instance-00000045 to attach disk [datastore2] volume-14e9fde5-2e48-4463-8391-cf95a9fa600a/volume-14e9fde5-2e48-4463-8391-cf95a9fa600a.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 892.951958] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53aadc70-903c-48ad-a053-c98bdff2a38e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.970812] env[61867]: DEBUG oslo_vmware.api [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 892.970812] env[61867]: value = "task-1276773" [ 892.970812] env[61867]: _type = "Task" [ 892.970812] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.984328] env[61867]: DEBUG oslo_vmware.api [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276773, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.987093] env[61867]: DEBUG oslo_concurrency.lockutils [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] Releasing lock "refresh_cache-adc2732d-2a10-40ce-bb90-ed0762a36614" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.987381] env[61867]: DEBUG nova.compute.manager [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Received event network-vif-plugged-65189f98-8c73-47e1-94cd-60c94c3a65d7 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.987574] env[61867]: DEBUG oslo_concurrency.lockutils [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] Acquiring lock "ca2f6d18-f773-4875-83de-2f2be912c2f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.987792] env[61867]: DEBUG oslo_concurrency.lockutils [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] Lock "ca2f6d18-f773-4875-83de-2f2be912c2f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.987955] env[61867]: DEBUG oslo_concurrency.lockutils [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] Lock "ca2f6d18-f773-4875-83de-2f2be912c2f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.988145] env[61867]: DEBUG nova.compute.manager [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] No waiting events found dispatching network-vif-plugged-65189f98-8c73-47e1-94cd-60c94c3a65d7 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 892.988320] env[61867]: WARNING nova.compute.manager [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Received unexpected event network-vif-plugged-65189f98-8c73-47e1-94cd-60c94c3a65d7 for instance with vm_state active and task_state None. [ 892.988490] env[61867]: DEBUG nova.compute.manager [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Received event network-changed-65189f98-8c73-47e1-94cd-60c94c3a65d7 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.988646] env[61867]: DEBUG nova.compute.manager [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Refreshing instance network info cache due to event network-changed-65189f98-8c73-47e1-94cd-60c94c3a65d7. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 892.988817] env[61867]: DEBUG oslo_concurrency.lockutils [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] Acquiring lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.005833] env[61867]: INFO nova.compute.manager [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] instance snapshotting [ 893.012033] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02997219-feaf-47d0-9656-7a827d467ce7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.034636] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e117b0-e8a3-4b5f-9737-68be51dae603 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.157800] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276772, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50378} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.158590] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] adc2732d-2a10-40ce-bb90-ed0762a36614/adc2732d-2a10-40ce-bb90-ed0762a36614.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 893.158748] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.158983] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-881b7245-3d60-45d3-812b-40794728f737 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.166259] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 893.166259] env[61867]: value = "task-1276774" [ 893.166259] env[61867]: _type = "Task" [ 893.166259] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.177813] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276774, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.339129] env[61867]: DEBUG nova.network.neutron [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Updating instance_info_cache with network_info: [{"id": "943c04a5-7eba-46db-ab4f-93f7f84be14e", "address": "fa:16:3e:56:de:39", "network": {"id": "7635ca79-1de7-4ba3-b347-1e0b66d7eb56", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1574772814-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8e9d6e6ceac4520add365429c4eb4d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d0c6fd7-3cc9-4818-9475-8f15900394cc", "external-id": "nsx-vlan-transportzone-317", "segmentation_id": 317, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap943c04a5-7e", "ovs_interfaceid": "943c04a5-7eba-46db-ab4f-93f7f84be14e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.342311] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.342950] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.343135] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.343412] env[61867]: DEBUG oslo_concurrency.lockutils [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] Acquired lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.343598] env[61867]: DEBUG nova.network.neutron [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Refreshing network info cache for port 65189f98-8c73-47e1-94cd-60c94c3a65d7 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 893.345306] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b4d09a-4127-4a1d-9184-6e7d7d08f0e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.364453] env[61867]: DEBUG nova.scheduler.client.report [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.367792] env[61867]: DEBUG nova.virt.hardware [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.368060] env[61867]: DEBUG nova.virt.hardware [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.368230] env[61867]: DEBUG nova.virt.hardware [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.368433] env[61867]: DEBUG nova.virt.hardware [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.368590] env[61867]: DEBUG nova.virt.hardware [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.368979] env[61867]: DEBUG nova.virt.hardware [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.368979] env[61867]: DEBUG nova.virt.hardware [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.369160] env[61867]: DEBUG nova.virt.hardware [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.369330] env[61867]: DEBUG nova.virt.hardware [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.369500] env[61867]: DEBUG nova.virt.hardware [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.369678] env[61867]: DEBUG nova.virt.hardware [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.376174] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Reconfiguring VM to attach interface {{(pid=61867) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 893.378248] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.827s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.378706] env[61867]: DEBUG nova.compute.manager [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 893.380989] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ebf5d3ce-5061-4061-abf1-1e549ec63146 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.393793] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.990s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.395275] env[61867]: INFO nova.compute.claims [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 893.406602] env[61867]: DEBUG oslo_vmware.api [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 893.406602] env[61867]: value = "task-1276775" [ 893.406602] env[61867]: _type = "Task" [ 893.406602] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.417158] env[61867]: DEBUG oslo_vmware.api [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276775, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.482248] env[61867]: DEBUG oslo_vmware.api [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276773, 'name': ReconfigVM_Task, 'duration_secs': 0.180501} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.482687] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274373', 'volume_id': '14e9fde5-2e48-4463-8391-cf95a9fa600a', 'name': 'volume-14e9fde5-2e48-4463-8391-cf95a9fa600a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6dda389d-a8c5-4e0e-87a5-4065e24c034e', 'attached_at': '', 'detached_at': '', 'volume_id': '14e9fde5-2e48-4463-8391-cf95a9fa600a', 'serial': '14e9fde5-2e48-4463-8391-cf95a9fa600a'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 893.546867] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Creating Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 893.547230] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-368d7ae6-2df3-4d7a-b031-89ba34e0a1dc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.561145] env[61867]: DEBUG oslo_vmware.api [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 893.561145] env[61867]: value = "task-1276776" [ 893.561145] env[61867]: _type = "Task" [ 893.561145] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.569977] env[61867]: DEBUG oslo_vmware.api [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276776, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.623610] env[61867]: DEBUG nova.compute.manager [req-dd99f75d-f13a-4ab1-ab40-3d7c33d84fca req-c43bc19e-4267-48c7-8c13-15d5f37d0ed1 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Received event network-vif-plugged-943c04a5-7eba-46db-ab4f-93f7f84be14e {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 893.624014] env[61867]: DEBUG oslo_concurrency.lockutils [req-dd99f75d-f13a-4ab1-ab40-3d7c33d84fca req-c43bc19e-4267-48c7-8c13-15d5f37d0ed1 service nova] Acquiring lock "1df8427c-e75d-4b60-a92a-b5ba76b67081-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.624121] env[61867]: DEBUG oslo_concurrency.lockutils [req-dd99f75d-f13a-4ab1-ab40-3d7c33d84fca req-c43bc19e-4267-48c7-8c13-15d5f37d0ed1 service nova] Lock "1df8427c-e75d-4b60-a92a-b5ba76b67081-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.624312] env[61867]: DEBUG oslo_concurrency.lockutils [req-dd99f75d-f13a-4ab1-ab40-3d7c33d84fca req-c43bc19e-4267-48c7-8c13-15d5f37d0ed1 service nova] Lock "1df8427c-e75d-4b60-a92a-b5ba76b67081-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.624600] env[61867]: DEBUG nova.compute.manager [req-dd99f75d-f13a-4ab1-ab40-3d7c33d84fca req-c43bc19e-4267-48c7-8c13-15d5f37d0ed1 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] No waiting events found dispatching network-vif-plugged-943c04a5-7eba-46db-ab4f-93f7f84be14e {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 893.624907] env[61867]: WARNING nova.compute.manager [req-dd99f75d-f13a-4ab1-ab40-3d7c33d84fca req-c43bc19e-4267-48c7-8c13-15d5f37d0ed1 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Received unexpected event network-vif-plugged-943c04a5-7eba-46db-ab4f-93f7f84be14e for instance with vm_state shelved_offloaded and task_state spawning. [ 893.625129] env[61867]: DEBUG nova.compute.manager [req-dd99f75d-f13a-4ab1-ab40-3d7c33d84fca req-c43bc19e-4267-48c7-8c13-15d5f37d0ed1 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Received event network-changed-943c04a5-7eba-46db-ab4f-93f7f84be14e {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 893.625314] env[61867]: DEBUG nova.compute.manager [req-dd99f75d-f13a-4ab1-ab40-3d7c33d84fca req-c43bc19e-4267-48c7-8c13-15d5f37d0ed1 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Refreshing instance network info cache due to event network-changed-943c04a5-7eba-46db-ab4f-93f7f84be14e. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 893.625496] env[61867]: DEBUG oslo_concurrency.lockutils [req-dd99f75d-f13a-4ab1-ab40-3d7c33d84fca req-c43bc19e-4267-48c7-8c13-15d5f37d0ed1 service nova] Acquiring lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.677049] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276774, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074817} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.677569] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.679047] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e668177-edf5-4e77-86ab-33c9de10e9b2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.704509] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] adc2732d-2a10-40ce-bb90-ed0762a36614/adc2732d-2a10-40ce-bb90-ed0762a36614.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.704831] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f329a244-62e5-4791-b7bf-948b67470ce9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.727700] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 893.727700] env[61867]: value = "task-1276777" [ 893.727700] env[61867]: _type = "Task" [ 893.727700] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.737229] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276777, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.821640] env[61867]: DEBUG oslo_vmware.rw_handles [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c96c07-5cfb-cbcb-19e0-6f10d3b78187/disk-0.vmdk. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 893.822474] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d06c1e1-2ced-43d4-b9e8-cf765c58f4e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.829537] env[61867]: DEBUG oslo_vmware.rw_handles [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c96c07-5cfb-cbcb-19e0-6f10d3b78187/disk-0.vmdk is in state: ready. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 893.829723] env[61867]: ERROR oslo_vmware.rw_handles [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c96c07-5cfb-cbcb-19e0-6f10d3b78187/disk-0.vmdk due to incomplete transfer. [ 893.829960] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-15221781-10ae-4c31-9aef-fc273508e0e5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.837640] env[61867]: DEBUG oslo_vmware.rw_handles [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c96c07-5cfb-cbcb-19e0-6f10d3b78187/disk-0.vmdk. {{(pid=61867) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 893.837893] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Uploaded image 928c7641-2c77-4753-814a-9eb159d83adf to the Glance image server {{(pid=61867) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 893.840368] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Destroying the VM {{(pid=61867) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 893.840657] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f0ba4cc0-a804-4251-bcf1-b46348635d4d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.842667] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Releasing lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.844993] env[61867]: DEBUG oslo_concurrency.lockutils [req-dd99f75d-f13a-4ab1-ab40-3d7c33d84fca req-c43bc19e-4267-48c7-8c13-15d5f37d0ed1 service nova] Acquired lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.845249] env[61867]: DEBUG nova.network.neutron [req-dd99f75d-f13a-4ab1-ab40-3d7c33d84fca req-c43bc19e-4267-48c7-8c13-15d5f37d0ed1 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Refreshing network info cache for port 943c04a5-7eba-46db-ab4f-93f7f84be14e {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 893.853988] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 893.853988] env[61867]: value = "task-1276778" [ 893.853988] env[61867]: _type = "Task" [ 893.853988] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.867935] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276778, 'name': Destroy_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.875419] env[61867]: DEBUG nova.virt.hardware [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='30c2bab05c29dc014ae095e060eb7b27',container_format='bare',created_at=2024-10-12T12:49:28Z,direct_url=,disk_format='vmdk',id=444eb1e0-4f45-40cf-b139-01bd39317edd,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1738941546-shelved',owner='c8e9d6e6ceac4520add365429c4eb4d2',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2024-10-12T12:49:45Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.875673] env[61867]: DEBUG nova.virt.hardware [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.875837] env[61867]: DEBUG nova.virt.hardware [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.876035] env[61867]: DEBUG nova.virt.hardware [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.876194] env[61867]: DEBUG nova.virt.hardware [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.876364] env[61867]: DEBUG nova.virt.hardware [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.876600] env[61867]: DEBUG nova.virt.hardware [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.876780] env[61867]: DEBUG nova.virt.hardware [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.876956] env[61867]: DEBUG nova.virt.hardware [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.880897] env[61867]: DEBUG nova.virt.hardware [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.881144] env[61867]: DEBUG nova.virt.hardware [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.882293] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f504658-5768-4ab8-accf-3a77d09a0f21 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.890490] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6908f761-a39f-47ff-8fbb-4f3ff15e8fba {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.906039] env[61867]: DEBUG nova.compute.utils [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 893.910341] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:56:de:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d0c6fd7-3cc9-4818-9475-8f15900394cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '943c04a5-7eba-46db-ab4f-93f7f84be14e', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 893.918775] env[61867]: DEBUG oslo.service.loopingcall [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 893.922503] env[61867]: DEBUG nova.compute.manager [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 893.922678] env[61867]: DEBUG nova.network.neutron [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 893.924366] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 893.930914] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c4385093-a1be-4b82-a686-a61856ede14f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.953872] env[61867]: DEBUG oslo_vmware.api [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276775, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.955235] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 893.955235] env[61867]: value = "task-1276779" [ 893.955235] env[61867]: _type = "Task" [ 893.955235] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.966028] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276779, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.996773] env[61867]: DEBUG nova.policy [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a013ca682054d16a6702230a22c180f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c28a8e133444440eb83dc6a848ef591d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 894.072026] env[61867]: DEBUG oslo_vmware.api [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276776, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.241154] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276777, 'name': ReconfigVM_Task, 'duration_secs': 0.273655} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.241555] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Reconfigured VM instance instance-00000052 to attach disk [datastore2] adc2732d-2a10-40ce-bb90-ed0762a36614/adc2732d-2a10-40ce-bb90-ed0762a36614.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.242340] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6b286190-b7fc-4c88-9351-152b5f87ebdf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.249502] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 894.249502] env[61867]: value = "task-1276780" [ 894.249502] env[61867]: _type = "Task" [ 894.249502] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.260406] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276780, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.325959] env[61867]: DEBUG nova.network.neutron [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Successfully created port: bad6d5ef-c6d0-448d-8a9b-01003e9872c7 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 894.365516] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276778, 'name': Destroy_Task} progress is 33%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.367762] env[61867]: DEBUG nova.network.neutron [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Updated VIF entry in instance network info cache for port 65189f98-8c73-47e1-94cd-60c94c3a65d7. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 894.368223] env[61867]: DEBUG nova.network.neutron [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Updating instance_info_cache with network_info: [{"id": "af89c243-c119-47c9-9eda-60c9be2b8c51", "address": "fa:16:3e:4d:1f:76", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf89c243-c1", "ovs_interfaceid": "af89c243-c119-47c9-9eda-60c9be2b8c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "65189f98-8c73-47e1-94cd-60c94c3a65d7", "address": "fa:16:3e:f2:05:68", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65189f98-8c", "ovs_interfaceid": "65189f98-8c73-47e1-94cd-60c94c3a65d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.408410] env[61867]: DEBUG nova.compute.manager [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 894.437276] env[61867]: DEBUG oslo_vmware.api [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276775, 'name': ReconfigVM_Task, 'duration_secs': 1.024738} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.437902] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.438362] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Reconfigured VM to attach interface {{(pid=61867) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 894.484167] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276779, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.526873] env[61867]: DEBUG nova.objects.instance [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lazy-loading 'flavor' on Instance uuid 6dda389d-a8c5-4e0e-87a5-4065e24c034e {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.579159] env[61867]: DEBUG oslo_vmware.api [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276776, 'name': CreateSnapshot_Task, 'duration_secs': 0.921839} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.579508] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Created Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 894.581325] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58aa8236-5ad1-430d-9195-21d7487c6e1d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.694412] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a614fc1-6e49-4c37-aa50-046594149a0d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.702854] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d9086e-daf9-418c-be07-adae70461549 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.738732] env[61867]: DEBUG nova.network.neutron [req-dd99f75d-f13a-4ab1-ab40-3d7c33d84fca req-c43bc19e-4267-48c7-8c13-15d5f37d0ed1 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Updated VIF entry in instance network info cache for port 943c04a5-7eba-46db-ab4f-93f7f84be14e. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 894.739227] env[61867]: DEBUG nova.network.neutron [req-dd99f75d-f13a-4ab1-ab40-3d7c33d84fca req-c43bc19e-4267-48c7-8c13-15d5f37d0ed1 service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Updating instance_info_cache with network_info: [{"id": "943c04a5-7eba-46db-ab4f-93f7f84be14e", "address": "fa:16:3e:56:de:39", "network": {"id": "7635ca79-1de7-4ba3-b347-1e0b66d7eb56", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1574772814-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8e9d6e6ceac4520add365429c4eb4d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d0c6fd7-3cc9-4818-9475-8f15900394cc", "external-id": "nsx-vlan-transportzone-317", "segmentation_id": 317, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap943c04a5-7e", "ovs_interfaceid": "943c04a5-7eba-46db-ab4f-93f7f84be14e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.740929] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19004e2e-b3ea-474f-b1be-bdcd9976c318 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.755019] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f14e500-e710-4b4c-9885-8b0a4a001254 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.765805] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276780, 'name': Rename_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.775255] env[61867]: DEBUG nova.compute.provider_tree [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.865671] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276778, 'name': Destroy_Task, 'duration_secs': 0.870425} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.866186] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Destroyed the VM [ 894.866275] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Deleting Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 894.866457] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9af4f8b0-1bbb-4f8b-bb4d-151252eeba72 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.870875] env[61867]: DEBUG oslo_concurrency.lockutils [req-03a337cd-4f82-422f-9b6f-01d5e98682c4 req-0e539735-1839-481d-90ef-32bd9db8d9bb service nova] Releasing lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.873496] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 894.873496] env[61867]: value = "task-1276781" [ 894.873496] env[61867]: _type = "Task" [ 894.873496] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.882712] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276781, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.947337] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e8828296-345f-41ed-8069-30a13c731bb4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "interface-ca2f6d18-f773-4875-83de-2f2be912c2f8-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.758s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.967659] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276779, 'name': CreateVM_Task, 'duration_secs': 0.62979} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.968564] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 894.969771] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/444eb1e0-4f45-40cf-b139-01bd39317edd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.969771] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquired lock "[datastore1] devstack-image-cache_base/444eb1e0-4f45-40cf-b139-01bd39317edd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.970069] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/444eb1e0-4f45-40cf-b139-01bd39317edd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 894.970431] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2a951fb-dd0c-4444-be6a-2df509206473 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.977571] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 894.977571] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52588085-95b5-c847-e8ba-0febbb969976" [ 894.977571] env[61867]: _type = "Task" [ 894.977571] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.987112] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52588085-95b5-c847-e8ba-0febbb969976, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.032933] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbb50983-eadb-4fd4-90f2-d8b0d600dc64 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.292s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.106914] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Creating linked-clone VM from snapshot {{(pid=61867) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 895.106914] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7460d8bc-6ebe-4edf-88d2-f70681aeafbe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.116770] env[61867]: DEBUG oslo_vmware.api [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 895.116770] env[61867]: value = "task-1276782" [ 895.116770] env[61867]: _type = "Task" [ 895.116770] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.125957] env[61867]: DEBUG oslo_vmware.api [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276782, 'name': CloneVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.245563] env[61867]: DEBUG oslo_concurrency.lockutils [req-dd99f75d-f13a-4ab1-ab40-3d7c33d84fca req-c43bc19e-4267-48c7-8c13-15d5f37d0ed1 service nova] Releasing lock "refresh_cache-1df8427c-e75d-4b60-a92a-b5ba76b67081" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.270346] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276780, 'name': Rename_Task, 'duration_secs': 0.566825} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.270801] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 895.271250] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-57d4b3e7-52c6-4acb-b78a-8b8ebdddc4a7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.282124] env[61867]: DEBUG nova.scheduler.client.report [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.283505] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 895.283505] env[61867]: value = "task-1276783" [ 895.283505] env[61867]: _type = "Task" [ 895.283505] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.294849] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276783, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.389314] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276781, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.422282] env[61867]: DEBUG nova.compute.manager [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 895.453500] env[61867]: DEBUG nova.virt.hardware [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 895.453817] env[61867]: DEBUG nova.virt.hardware [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 895.453990] env[61867]: DEBUG nova.virt.hardware [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 895.454299] env[61867]: DEBUG nova.virt.hardware [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 895.454491] env[61867]: DEBUG nova.virt.hardware [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 895.454607] env[61867]: DEBUG nova.virt.hardware [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 895.454919] env[61867]: DEBUG nova.virt.hardware [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 895.455539] env[61867]: DEBUG nova.virt.hardware [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 895.455776] env[61867]: DEBUG nova.virt.hardware [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 895.456051] env[61867]: DEBUG nova.virt.hardware [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 895.456159] env[61867]: DEBUG nova.virt.hardware [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 895.457876] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90c5078-cd76-4507-a415-17cd0db45dcf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.469965] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4da46ba-fa29-4589-af11-5eadc0885466 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.495646] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Releasing lock "[datastore1] devstack-image-cache_base/444eb1e0-4f45-40cf-b139-01bd39317edd" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.495879] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Processing image 444eb1e0-4f45-40cf-b139-01bd39317edd {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.496104] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/444eb1e0-4f45-40cf-b139-01bd39317edd/444eb1e0-4f45-40cf-b139-01bd39317edd.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.496705] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquired lock "[datastore1] devstack-image-cache_base/444eb1e0-4f45-40cf-b139-01bd39317edd/444eb1e0-4f45-40cf-b139-01bd39317edd.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.496705] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.496705] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d04a1fad-644e-4d67-8442-df17e6d7c858 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.507619] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.507873] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 895.508669] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93b91769-fd94-4f79-be24-5f5676a4cd11 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.515827] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 895.515827] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f9451c-28ab-5d51-d344-cc64c3a530a0" [ 895.515827] env[61867]: _type = "Task" [ 895.515827] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.525633] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f9451c-28ab-5d51-d344-cc64c3a530a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.631659] env[61867]: DEBUG oslo_vmware.api [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276782, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.789022] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.789022] env[61867]: DEBUG nova.compute.manager [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 895.792014] env[61867]: DEBUG oslo_concurrency.lockutils [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.976s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.792289] env[61867]: DEBUG nova.objects.instance [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lazy-loading 'resources' on Instance uuid 07fdd75c-34ab-45e5-a98b-aced2caa6cb5 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.808142] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276783, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.886745] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276781, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.025178] env[61867]: DEBUG nova.compute.manager [req-bf0161d2-be9d-402c-891c-f5f193b1bf03 req-c811931a-1e9d-4cc9-a1e4-c92929a9d41b service nova] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Received event network-vif-plugged-bad6d5ef-c6d0-448d-8a9b-01003e9872c7 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.025178] env[61867]: DEBUG oslo_concurrency.lockutils [req-bf0161d2-be9d-402c-891c-f5f193b1bf03 req-c811931a-1e9d-4cc9-a1e4-c92929a9d41b service nova] Acquiring lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.025178] env[61867]: DEBUG oslo_concurrency.lockutils [req-bf0161d2-be9d-402c-891c-f5f193b1bf03 req-c811931a-1e9d-4cc9-a1e4-c92929a9d41b service nova] Lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.025178] env[61867]: DEBUG oslo_concurrency.lockutils [req-bf0161d2-be9d-402c-891c-f5f193b1bf03 req-c811931a-1e9d-4cc9-a1e4-c92929a9d41b service nova] Lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.025178] env[61867]: DEBUG nova.compute.manager [req-bf0161d2-be9d-402c-891c-f5f193b1bf03 req-c811931a-1e9d-4cc9-a1e4-c92929a9d41b service nova] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] No waiting events found dispatching network-vif-plugged-bad6d5ef-c6d0-448d-8a9b-01003e9872c7 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 896.025706] env[61867]: WARNING nova.compute.manager [req-bf0161d2-be9d-402c-891c-f5f193b1bf03 req-c811931a-1e9d-4cc9-a1e4-c92929a9d41b service nova] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Received unexpected event network-vif-plugged-bad6d5ef-c6d0-448d-8a9b-01003e9872c7 for instance with vm_state building and task_state spawning. [ 896.035196] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Preparing fetch location {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 896.036673] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Fetch image to [datastore1] OSTACK_IMG_a6ec9839-2551-4776-a3fb-5540d4202dab/OSTACK_IMG_a6ec9839-2551-4776-a3fb-5540d4202dab.vmdk {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 896.037590] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Downloading stream optimized image 444eb1e0-4f45-40cf-b139-01bd39317edd to [datastore1] OSTACK_IMG_a6ec9839-2551-4776-a3fb-5540d4202dab/OSTACK_IMG_a6ec9839-2551-4776-a3fb-5540d4202dab.vmdk on the data store datastore1 as vApp {{(pid=61867) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 896.037590] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Downloading image file data 444eb1e0-4f45-40cf-b139-01bd39317edd to the ESX as VM named 'OSTACK_IMG_a6ec9839-2551-4776-a3fb-5540d4202dab' {{(pid=61867) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 896.087196] env[61867]: DEBUG nova.network.neutron [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Successfully updated port: bad6d5ef-c6d0-448d-8a9b-01003e9872c7 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 896.131172] env[61867]: DEBUG oslo_vmware.api [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276782, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.141398] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 896.141398] env[61867]: value = "resgroup-9" [ 896.141398] env[61867]: _type = "ResourcePool" [ 896.141398] env[61867]: }. {{(pid=61867) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 896.142089] env[61867]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-35ebfab5-4240-446b-9dde-d993c07990dc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.179110] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lease: (returnval){ [ 896.179110] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52c81a0f-d744-3a1f-ed08-54a943c1d826" [ 896.179110] env[61867]: _type = "HttpNfcLease" [ 896.179110] env[61867]: } obtained for vApp import into resource pool (val){ [ 896.179110] env[61867]: value = "resgroup-9" [ 896.179110] env[61867]: _type = "ResourcePool" [ 896.179110] env[61867]: }. {{(pid=61867) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 896.179480] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the lease: (returnval){ [ 896.179480] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52c81a0f-d744-3a1f-ed08-54a943c1d826" [ 896.179480] env[61867]: _type = "HttpNfcLease" [ 896.179480] env[61867]: } to be ready. {{(pid=61867) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 896.191030] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 896.191030] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52c81a0f-d744-3a1f-ed08-54a943c1d826" [ 896.191030] env[61867]: _type = "HttpNfcLease" [ 896.191030] env[61867]: } is initializing. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 896.294057] env[61867]: DEBUG nova.compute.utils [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 896.300420] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.300862] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.005s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.313022] env[61867]: DEBUG nova.compute.manager [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 896.315815] env[61867]: DEBUG nova.network.neutron [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 896.324618] env[61867]: INFO nova.compute.manager [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Detaching volume b788827a-5734-4c3e-a38d-82845c30c290 [ 896.343461] env[61867]: DEBUG oslo_vmware.api [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276783, 'name': PowerOnVM_Task, 'duration_secs': 0.598853} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.345516] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 896.345516] env[61867]: INFO nova.compute.manager [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Took 7.12 seconds to spawn the instance on the hypervisor. [ 896.345516] env[61867]: DEBUG nova.compute.manager [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.346412] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c69f234f-3623-45e4-a0cd-b8b00b4a542e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.383980] env[61867]: DEBUG nova.policy [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12ed008ec3204102b2b08e61ed24f418', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '47c5b5b783ea445eb1c02fb728b1fff2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 896.389705] env[61867]: DEBUG oslo_vmware.api [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276781, 'name': RemoveSnapshot_Task, 'duration_secs': 1.088338} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.389705] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Deleted Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 896.389913] env[61867]: INFO nova.compute.manager [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Took 16.60 seconds to snapshot the instance on the hypervisor. [ 896.402827] env[61867]: INFO nova.virt.block_device [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Attempting to driver detach volume b788827a-5734-4c3e-a38d-82845c30c290 from mountpoint /dev/sdb [ 896.402827] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Volume detach. Driver type: vmdk {{(pid=61867) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 896.402827] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274369', 'volume_id': 'b788827a-5734-4c3e-a38d-82845c30c290', 'name': 'volume-b788827a-5734-4c3e-a38d-82845c30c290', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6dda389d-a8c5-4e0e-87a5-4065e24c034e', 'attached_at': '', 'detached_at': '', 'volume_id': 'b788827a-5734-4c3e-a38d-82845c30c290', 'serial': 'b788827a-5734-4c3e-a38d-82845c30c290'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 896.403713] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50fc8167-b775-4ee2-aba3-9c5730d3c782 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.438047] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ae803b-a8c2-4ca7-853f-444731690e2d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.452451] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03a78222-d551-49bd-bcdf-5952279050cd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.487897] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b72b0c-cf7a-4399-b5de-c83b02a33525 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.506942] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] The volume has not been displaced from its original location: [datastore1] volume-b788827a-5734-4c3e-a38d-82845c30c290/volume-b788827a-5734-4c3e-a38d-82845c30c290.vmdk. No consolidation needed. {{(pid=61867) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 896.512539] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Reconfiguring VM instance instance-00000045 to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 896.515651] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-213048ce-6825-4b7b-b5ec-648ce941d2ea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.535754] env[61867]: DEBUG oslo_vmware.api [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 896.535754] env[61867]: value = "task-1276785" [ 896.535754] env[61867]: _type = "Task" [ 896.535754] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.545513] env[61867]: DEBUG oslo_vmware.api [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276785, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.590480] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "refresh_cache-9e1cbfa9-28e9-4bca-adfd-78bf25428106" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.590673] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "refresh_cache-9e1cbfa9-28e9-4bca-adfd-78bf25428106" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.590786] env[61867]: DEBUG nova.network.neutron [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 896.632423] env[61867]: DEBUG oslo_vmware.api [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276782, 'name': CloneVM_Task, 'duration_secs': 1.272285} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.632619] env[61867]: INFO nova.virt.vmwareapi.vmops [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Created linked-clone VM from snapshot [ 896.633877] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63819c06-58ff-450c-bcd6-14daade7344c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.642486] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Uploading image bf98b492-69a7-4ae7-b48d-543f17ad8b20 {{(pid=61867) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 896.653899] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Destroying the VM {{(pid=61867) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 896.654209] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9b771f53-01f2-4644-8088-7e9c8cf37016 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.665858] env[61867]: DEBUG oslo_vmware.api [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 896.665858] env[61867]: value = "task-1276786" [ 896.665858] env[61867]: _type = "Task" [ 896.665858] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.679666] env[61867]: DEBUG oslo_vmware.api [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276786, 'name': Destroy_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.685599] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a652a46a-e7f5-4b6e-aadf-e7b7d9e37201 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.692318] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 896.692318] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52c81a0f-d744-3a1f-ed08-54a943c1d826" [ 896.692318] env[61867]: _type = "HttpNfcLease" [ 896.692318] env[61867]: } is initializing. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 896.692877] env[61867]: DEBUG nova.network.neutron [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Successfully created port: 254a9643-f941-493e-8c87-2932a6cc00ce {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 896.697710] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a176fc-0445-4dfc-afbd-260338758470 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.729737] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-011cc6c3-b592-4c85-a519-c87884a80ff0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.738428] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c08c2d8c-5be8-4f33-9341-8c870f8a46e7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.756750] env[61867]: DEBUG nova.compute.provider_tree [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.804255] env[61867]: DEBUG nova.compute.manager [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 896.866205] env[61867]: INFO nova.compute.manager [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Took 19.39 seconds to build instance. [ 896.948806] env[61867]: DEBUG nova.compute.manager [None req-1a48900d-c101-4f0d-83e5-6d478cacef74 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Found 2 images (rotation: 2) {{(pid=61867) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 897.052017] env[61867]: DEBUG oslo_vmware.api [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276785, 'name': ReconfigVM_Task, 'duration_secs': 0.270057} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.052017] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Reconfigured VM instance instance-00000045 to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 897.054772] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5a929cf-7798-4d48-afbc-152dc775adba {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.075286] env[61867]: DEBUG oslo_vmware.api [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 897.075286] env[61867]: value = "task-1276787" [ 897.075286] env[61867]: _type = "Task" [ 897.075286] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.087213] env[61867]: DEBUG oslo_vmware.api [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276787, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.153444] env[61867]: DEBUG nova.network.neutron [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 897.179255] env[61867]: DEBUG oslo_vmware.api [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276786, 'name': Destroy_Task} progress is 33%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.190297] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 897.190297] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52c81a0f-d744-3a1f-ed08-54a943c1d826" [ 897.190297] env[61867]: _type = "HttpNfcLease" [ 897.190297] env[61867]: } is ready. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 897.190297] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 897.190297] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52c81a0f-d744-3a1f-ed08-54a943c1d826" [ 897.190297] env[61867]: _type = "HttpNfcLease" [ 897.190297] env[61867]: }. {{(pid=61867) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 897.190297] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871d282d-98c3-4fb4-b46b-8d5f883c627c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.199126] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5280debf-69eb-ffe2-d701-92babcf779d4/disk-0.vmdk from lease info. {{(pid=61867) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 897.199486] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5280debf-69eb-ffe2-d701-92babcf779d4/disk-0.vmdk. {{(pid=61867) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 897.271909] env[61867]: DEBUG nova.scheduler.client.report [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.282242] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c3181d68-cbd4-41b0-aec5-d4029363be20 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.370225] env[61867]: DEBUG oslo_concurrency.lockutils [None req-15c64eab-f0f9-4b40-bae2-7f56d45756df tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "adc2732d-2a10-40ce-bb90-ed0762a36614" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.903s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.414350] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "interface-ca2f6d18-f773-4875-83de-2f2be912c2f8-a3bb848c-f036-4379-9e73-e7089972da31" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.415011] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "interface-ca2f6d18-f773-4875-83de-2f2be912c2f8-a3bb848c-f036-4379-9e73-e7089972da31" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.415152] env[61867]: DEBUG nova.objects.instance [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lazy-loading 'flavor' on Instance uuid ca2f6d18-f773-4875-83de-2f2be912c2f8 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.450576] env[61867]: DEBUG nova.network.neutron [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Updating instance_info_cache with network_info: [{"id": "bad6d5ef-c6d0-448d-8a9b-01003e9872c7", "address": "fa:16:3e:45:db:29", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbad6d5ef-c6", "ovs_interfaceid": "bad6d5ef-c6d0-448d-8a9b-01003e9872c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.592070] env[61867]: DEBUG oslo_vmware.api [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276787, 'name': ReconfigVM_Task, 'duration_secs': 0.152806} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.594184] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274369', 'volume_id': 'b788827a-5734-4c3e-a38d-82845c30c290', 'name': 'volume-b788827a-5734-4c3e-a38d-82845c30c290', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6dda389d-a8c5-4e0e-87a5-4065e24c034e', 'attached_at': '', 'detached_at': '', 'volume_id': 'b788827a-5734-4c3e-a38d-82845c30c290', 'serial': 'b788827a-5734-4c3e-a38d-82845c30c290'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 897.699987] env[61867]: DEBUG oslo_vmware.api [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276786, 'name': Destroy_Task, 'duration_secs': 0.617334} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.702549] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Destroyed the VM [ 897.702880] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Deleting Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 897.703678] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-85e0892e-2d9e-4584-bc23-ccd45b7c423e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.711791] env[61867]: DEBUG oslo_vmware.api [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 897.711791] env[61867]: value = "task-1276788" [ 897.711791] env[61867]: _type = "Task" [ 897.711791] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.727868] env[61867]: DEBUG oslo_vmware.api [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276788, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.753219] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "816ca796-d8ed-4843-9b1f-f169f48ff047" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.754044] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "816ca796-d8ed-4843-9b1f-f169f48ff047" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.780332] env[61867]: DEBUG oslo_concurrency.lockutils [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.988s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.802047] env[61867]: INFO nova.scheduler.client.report [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleted allocations for instance 07fdd75c-34ab-45e5-a98b-aced2caa6cb5 [ 897.815814] env[61867]: DEBUG nova.compute.manager [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 897.844755] env[61867]: DEBUG nova.virt.hardware [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 897.846423] env[61867]: DEBUG nova.virt.hardware [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 897.846423] env[61867]: DEBUG nova.virt.hardware [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 897.846423] env[61867]: DEBUG nova.virt.hardware [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 897.846423] env[61867]: DEBUG nova.virt.hardware [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 897.846423] env[61867]: DEBUG nova.virt.hardware [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 897.846423] env[61867]: DEBUG nova.virt.hardware [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 897.846423] env[61867]: DEBUG nova.virt.hardware [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 897.846678] env[61867]: DEBUG nova.virt.hardware [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 897.848303] env[61867]: DEBUG nova.virt.hardware [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 897.848303] env[61867]: DEBUG nova.virt.hardware [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 897.850020] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45649de5-b5fb-4eda-b1a8-81fd8bb38994 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.860027] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47210f6-a4ef-4a9c-ad7a-8d8f129d7a2e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.953336] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "refresh_cache-9e1cbfa9-28e9-4bca-adfd-78bf25428106" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.953835] env[61867]: DEBUG nova.compute.manager [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Instance network_info: |[{"id": "bad6d5ef-c6d0-448d-8a9b-01003e9872c7", "address": "fa:16:3e:45:db:29", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbad6d5ef-c6", "ovs_interfaceid": "bad6d5ef-c6d0-448d-8a9b-01003e9872c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 897.954351] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:db:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74e6f6e0-95e6-4531-99e9-0e78350fb655', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bad6d5ef-c6d0-448d-8a9b-01003e9872c7', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.962735] env[61867]: DEBUG oslo.service.loopingcall [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.969278] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 897.969600] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-14c4111b-a767-440e-a058-3ff52b148a8f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.003066] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 898.003066] env[61867]: value = "task-1276789" [ 898.003066] env[61867]: _type = "Task" [ 898.003066] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.018899] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276789, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.105347] env[61867]: DEBUG nova.objects.instance [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lazy-loading 'pci_requests' on Instance uuid ca2f6d18-f773-4875-83de-2f2be912c2f8 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.158221] env[61867]: DEBUG nova.objects.instance [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lazy-loading 'flavor' on Instance uuid 6dda389d-a8c5-4e0e-87a5-4065e24c034e {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.222262] env[61867]: DEBUG oslo_vmware.api [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276788, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.263755] env[61867]: DEBUG nova.compute.manager [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 898.311974] env[61867]: DEBUG oslo_concurrency.lockutils [None req-35a71140-fed5-44f7-ae09-525617b1a573 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "07fdd75c-34ab-45e5-a98b-aced2caa6cb5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.426s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.439367] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "5ec9eb1f-1d24-41d9-b5c7-a478a427c563" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.439674] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "5ec9eb1f-1d24-41d9-b5c7-a478a427c563" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.456535] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Completed reading data from the image iterator. {{(pid=61867) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 898.456806] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5280debf-69eb-ffe2-d701-92babcf779d4/disk-0.vmdk. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 898.457621] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a046359-ace4-40d2-835b-50a81b8f552c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.467788] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5280debf-69eb-ffe2-d701-92babcf779d4/disk-0.vmdk is in state: ready. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 898.468023] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5280debf-69eb-ffe2-d701-92babcf779d4/disk-0.vmdk. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 898.468308] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-f8aefc99-ed36-4e33-8f64-44f3df4e8c33 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.514734] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276789, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.587189] env[61867]: DEBUG nova.network.neutron [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Successfully updated port: 254a9643-f941-493e-8c87-2932a6cc00ce {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 898.609040] env[61867]: DEBUG nova.objects.base [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61867) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 898.609303] env[61867]: DEBUG nova.network.neutron [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 898.678379] env[61867]: DEBUG nova.policy [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ea634246436422a9407f82692b69ede', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '86ced20eaf4740e298dc6f8ca5550c09', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 898.712975] env[61867]: DEBUG nova.compute.manager [req-9190e154-cd34-4c0c-9e8d-2689b112243b req-e9ed95ec-1243-4ede-bd7a-77f0b362128c service nova] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Received event network-changed-bad6d5ef-c6d0-448d-8a9b-01003e9872c7 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.713260] env[61867]: DEBUG nova.compute.manager [req-9190e154-cd34-4c0c-9e8d-2689b112243b req-e9ed95ec-1243-4ede-bd7a-77f0b362128c service nova] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Refreshing instance network info cache due to event network-changed-bad6d5ef-c6d0-448d-8a9b-01003e9872c7. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 898.713528] env[61867]: DEBUG oslo_concurrency.lockutils [req-9190e154-cd34-4c0c-9e8d-2689b112243b req-e9ed95ec-1243-4ede-bd7a-77f0b362128c service nova] Acquiring lock "refresh_cache-9e1cbfa9-28e9-4bca-adfd-78bf25428106" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.713697] env[61867]: DEBUG oslo_concurrency.lockutils [req-9190e154-cd34-4c0c-9e8d-2689b112243b req-e9ed95ec-1243-4ede-bd7a-77f0b362128c service nova] Acquired lock "refresh_cache-9e1cbfa9-28e9-4bca-adfd-78bf25428106" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.713867] env[61867]: DEBUG nova.network.neutron [req-9190e154-cd34-4c0c-9e8d-2689b112243b req-e9ed95ec-1243-4ede-bd7a-77f0b362128c service nova] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Refreshing network info cache for port bad6d5ef-c6d0-448d-8a9b-01003e9872c7 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 898.716807] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5280debf-69eb-ffe2-d701-92babcf779d4/disk-0.vmdk. {{(pid=61867) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 898.717038] env[61867]: INFO nova.virt.vmwareapi.images [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Downloaded image file data 444eb1e0-4f45-40cf-b139-01bd39317edd [ 898.721496] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8630c1b9-3b1e-41ca-b4d1-78f2a9567604 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.731866] env[61867]: DEBUG oslo_vmware.api [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276788, 'name': RemoveSnapshot_Task, 'duration_secs': 0.880279} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.744507] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Deleted Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 898.747617] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bce2ea7c-67b9-47bc-aa7d-343633167b8a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.779190] env[61867]: INFO nova.virt.vmwareapi.images [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] The imported VM was unregistered [ 898.781447] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Caching image {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 898.781684] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Creating directory with path [datastore1] devstack-image-cache_base/444eb1e0-4f45-40cf-b139-01bd39317edd {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.782597] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6f946733-28ed-4b9a-9dfc-0f7760f1e2bd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.791484] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.792014] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.793565] env[61867]: INFO nova.compute.claims [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 898.802041] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Created directory with path [datastore1] devstack-image-cache_base/444eb1e0-4f45-40cf-b139-01bd39317edd {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.802279] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_a6ec9839-2551-4776-a3fb-5540d4202dab/OSTACK_IMG_a6ec9839-2551-4776-a3fb-5540d4202dab.vmdk to [datastore1] devstack-image-cache_base/444eb1e0-4f45-40cf-b139-01bd39317edd/444eb1e0-4f45-40cf-b139-01bd39317edd.vmdk. {{(pid=61867) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 898.802568] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-2688dbde-18d5-4ead-9f5c-f30807a22d61 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.810412] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 898.810412] env[61867]: value = "task-1276791" [ 898.810412] env[61867]: _type = "Task" [ 898.810412] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.819012] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276791, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.945447] env[61867]: DEBUG nova.compute.manager [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 899.014898] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276789, 'name': CreateVM_Task, 'duration_secs': 0.521991} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.015124] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 899.015816] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.015988] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.016337] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 899.016608] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f863d81d-3365-451b-bd10-b93c916cee3d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.021644] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 899.021644] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52277a8d-f8ea-028e-5f08-eef7c8f4a83d" [ 899.021644] env[61867]: _type = "Task" [ 899.021644] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.030020] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52277a8d-f8ea-028e-5f08-eef7c8f4a83d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.089465] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "refresh_cache-705a32b6-67f9-42cc-b4d0-f6d1783c68b5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.089645] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "refresh_cache-705a32b6-67f9-42cc-b4d0-f6d1783c68b5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.089819] env[61867]: DEBUG nova.network.neutron [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 899.170025] env[61867]: DEBUG oslo_concurrency.lockutils [None req-dd5a4acb-c069-449c-84e6-197c807232d5 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 2.869s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.193410] env[61867]: DEBUG nova.compute.manager [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 899.194065] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b81209f-6ce5-4b92-bf4f-44aed488fae1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.251139] env[61867]: WARNING nova.compute.manager [None req-90eab34f-bc7e-4e45-8839-2c007f296a0b tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Image not found during snapshot: nova.exception.ImageNotFound: Image bf98b492-69a7-4ae7-b48d-543f17ad8b20 could not be found. [ 899.325063] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276791, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.438801] env[61867]: DEBUG nova.network.neutron [req-9190e154-cd34-4c0c-9e8d-2689b112243b req-e9ed95ec-1243-4ede-bd7a-77f0b362128c service nova] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Updated VIF entry in instance network info cache for port bad6d5ef-c6d0-448d-8a9b-01003e9872c7. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 899.439275] env[61867]: DEBUG nova.network.neutron [req-9190e154-cd34-4c0c-9e8d-2689b112243b req-e9ed95ec-1243-4ede-bd7a-77f0b362128c service nova] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Updating instance_info_cache with network_info: [{"id": "bad6d5ef-c6d0-448d-8a9b-01003e9872c7", "address": "fa:16:3e:45:db:29", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbad6d5ef-c6", "ovs_interfaceid": "bad6d5ef-c6d0-448d-8a9b-01003e9872c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.469127] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.535810] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52277a8d-f8ea-028e-5f08-eef7c8f4a83d, 'name': SearchDatastore_Task, 'duration_secs': 0.02392} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.536204] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.536482] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 899.536776] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.536934] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.537162] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 899.537490] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e9ee7eb0-6bf0-4e83-80dd-f39801334c0f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.552702] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 899.552963] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 899.553871] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eba28fe3-458d-4880-a595-f1dfe9e61763 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.562902] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 899.562902] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529dcafd-344b-a3cf-ec61-727244b9319b" [ 899.562902] env[61867]: _type = "Task" [ 899.562902] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.573720] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529dcafd-344b-a3cf-ec61-727244b9319b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.622861] env[61867]: DEBUG nova.network.neutron [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 899.706336] env[61867]: INFO nova.compute.manager [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] instance snapshotting [ 899.706813] env[61867]: DEBUG nova.objects.instance [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lazy-loading 'flavor' on Instance uuid abb41c0c-6d0d-4147-a4af-554ab7d9e921 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.765643] env[61867]: DEBUG nova.network.neutron [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Updating instance_info_cache with network_info: [{"id": "254a9643-f941-493e-8c87-2932a6cc00ce", "address": "fa:16:3e:85:01:b4", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap254a9643-f9", "ovs_interfaceid": "254a9643-f941-493e-8c87-2932a6cc00ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.824276] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276791, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.860088] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.860088] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.941749] env[61867]: DEBUG oslo_concurrency.lockutils [req-9190e154-cd34-4c0c-9e8d-2689b112243b req-e9ed95ec-1243-4ede-bd7a-77f0b362128c service nova] Releasing lock "refresh_cache-9e1cbfa9-28e9-4bca-adfd-78bf25428106" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.033519] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea543b1-618a-4565-8948-9033b61b0e69 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.044637] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a3e58c-3b88-437a-9240-6e500d44f322 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.083350] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516d5637-e26e-4ab8-b76e-f799262723fd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.096534] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fa0e8ee-ed2c-41c6-bac9-f1e0796ca1ea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.100789] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529dcafd-344b-a3cf-ec61-727244b9319b, 'name': SearchDatastore_Task, 'duration_secs': 0.054682} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.102116] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5afb7166-1d0e-4116-80de-a1cc9ab05c3e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.113339] env[61867]: DEBUG nova.compute.provider_tree [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.116182] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 900.116182] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52c91db4-9814-0369-c9bd-b3a49de98f06" [ 900.116182] env[61867]: _type = "Task" [ 900.116182] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.126343] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52c91db4-9814-0369-c9bd-b3a49de98f06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.215122] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35cdf5af-49c9-4eeb-8b69-bedda83e2fcd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.220526] env[61867]: DEBUG nova.network.neutron [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Successfully updated port: a3bb848c-f036-4379-9e73-e7089972da31 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 900.240528] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.240719] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.240937] env[61867]: DEBUG nova.network.neutron [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 900.243208] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4781aedf-58f6-444b-b6e7-8039659bd246 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.268482] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "refresh_cache-705a32b6-67f9-42cc-b4d0-f6d1783c68b5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.268810] env[61867]: DEBUG nova.compute.manager [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Instance network_info: |[{"id": "254a9643-f941-493e-8c87-2932a6cc00ce", "address": "fa:16:3e:85:01:b4", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap254a9643-f9", "ovs_interfaceid": "254a9643-f941-493e-8c87-2932a6cc00ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 900.269407] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:01:b4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f2020f39-42c4-4481-85c5-aaf03854b459', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '254a9643-f941-493e-8c87-2932a6cc00ce', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 900.277221] env[61867]: DEBUG oslo.service.loopingcall [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.277858] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 900.278311] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b7751e6e-b4b9-470d-8d34-437414dd9073 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.300342] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 900.300342] env[61867]: value = "task-1276792" [ 900.300342] env[61867]: _type = "Task" [ 900.300342] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.309685] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276792, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.320759] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276791, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.369260] env[61867]: INFO nova.compute.manager [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Detaching volume 14e9fde5-2e48-4463-8391-cf95a9fa600a [ 900.411885] env[61867]: INFO nova.virt.block_device [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Attempting to driver detach volume 14e9fde5-2e48-4463-8391-cf95a9fa600a from mountpoint /dev/sdc [ 900.412047] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Volume detach. Driver type: vmdk {{(pid=61867) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 900.412306] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274373', 'volume_id': '14e9fde5-2e48-4463-8391-cf95a9fa600a', 'name': 'volume-14e9fde5-2e48-4463-8391-cf95a9fa600a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6dda389d-a8c5-4e0e-87a5-4065e24c034e', 'attached_at': '', 'detached_at': '', 'volume_id': '14e9fde5-2e48-4463-8391-cf95a9fa600a', 'serial': '14e9fde5-2e48-4463-8391-cf95a9fa600a'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 900.413576] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2df6e982-9977-4b98-a5b2-06cb61a835fd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.446949] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a756d60-71ba-4fdb-ac4c-c108df765f5e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.455047] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1de24b-0f8d-4629-8b98-bf6b75f04735 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.476164] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6630940e-7f70-45e8-a127-4d50b181093f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.494253] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] The volume has not been displaced from its original location: [datastore2] volume-14e9fde5-2e48-4463-8391-cf95a9fa600a/volume-14e9fde5-2e48-4463-8391-cf95a9fa600a.vmdk. No consolidation needed. {{(pid=61867) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 900.499707] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Reconfiguring VM instance instance-00000045 to detach disk 2002 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 900.500129] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ff21880-cc9d-4a01-9a92-fcbf4e39148c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.519527] env[61867]: DEBUG oslo_vmware.api [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 900.519527] env[61867]: value = "task-1276793" [ 900.519527] env[61867]: _type = "Task" [ 900.519527] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.528145] env[61867]: DEBUG oslo_vmware.api [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276793, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.618797] env[61867]: DEBUG nova.scheduler.client.report [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 900.632137] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52c91db4-9814-0369-c9bd-b3a49de98f06, 'name': SearchDatastore_Task, 'duration_secs': 0.093959} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.632438] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.632714] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 9e1cbfa9-28e9-4bca-adfd-78bf25428106/9e1cbfa9-28e9-4bca-adfd-78bf25428106.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 900.632985] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fba5fcb5-7b9d-4bd2-b429-3c12c8f19124 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.639959] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 900.639959] env[61867]: value = "task-1276794" [ 900.639959] env[61867]: _type = "Task" [ 900.639959] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.649909] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276794, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.757403] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Creating Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 900.757796] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-18e5c3d8-c116-4383-824f-d968a76d95d7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.768391] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 900.768391] env[61867]: value = "task-1276795" [ 900.768391] env[61867]: _type = "Task" [ 900.768391] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.777163] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276795, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.785822] env[61867]: WARNING nova.network.neutron [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] 7a62cd67-4e7d-4952-9a24-053e35c1c3cb already exists in list: networks containing: ['7a62cd67-4e7d-4952-9a24-053e35c1c3cb']. ignoring it [ 900.786140] env[61867]: WARNING nova.network.neutron [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] 7a62cd67-4e7d-4952-9a24-053e35c1c3cb already exists in list: networks containing: ['7a62cd67-4e7d-4952-9a24-053e35c1c3cb']. ignoring it [ 900.810495] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276792, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.820679] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276791, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.874211] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "b8e561df-069e-4873-91c3-1bfe5f27de2d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.874482] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "b8e561df-069e-4873-91c3-1bfe5f27de2d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.874698] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "b8e561df-069e-4873-91c3-1bfe5f27de2d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.874885] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "b8e561df-069e-4873-91c3-1bfe5f27de2d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.875072] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "b8e561df-069e-4873-91c3-1bfe5f27de2d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.877217] env[61867]: INFO nova.compute.manager [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Terminating instance [ 900.879187] env[61867]: DEBUG nova.compute.manager [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 900.879389] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 900.880236] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-724bf479-63b8-4e3a-8964-77636ed6aece {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.888866] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 900.889087] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e48e3d6-c062-4a23-994e-aa53e6dad05f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.896347] env[61867]: DEBUG oslo_vmware.api [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 900.896347] env[61867]: value = "task-1276796" [ 900.896347] env[61867]: _type = "Task" [ 900.896347] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.904790] env[61867]: DEBUG oslo_vmware.api [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276796, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.951222] env[61867]: DEBUG nova.compute.manager [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Received event network-vif-plugged-254a9643-f941-493e-8c87-2932a6cc00ce {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.951530] env[61867]: DEBUG oslo_concurrency.lockutils [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] Acquiring lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.951751] env[61867]: DEBUG oslo_concurrency.lockutils [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] Lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.951924] env[61867]: DEBUG oslo_concurrency.lockutils [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] Lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.952109] env[61867]: DEBUG nova.compute.manager [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] No waiting events found dispatching network-vif-plugged-254a9643-f941-493e-8c87-2932a6cc00ce {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 900.952337] env[61867]: WARNING nova.compute.manager [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Received unexpected event network-vif-plugged-254a9643-f941-493e-8c87-2932a6cc00ce for instance with vm_state building and task_state spawning. [ 900.952592] env[61867]: DEBUG nova.compute.manager [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Received event network-changed-254a9643-f941-493e-8c87-2932a6cc00ce {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.952712] env[61867]: DEBUG nova.compute.manager [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Refreshing instance network info cache due to event network-changed-254a9643-f941-493e-8c87-2932a6cc00ce. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 900.952915] env[61867]: DEBUG oslo_concurrency.lockutils [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] Acquiring lock "refresh_cache-705a32b6-67f9-42cc-b4d0-f6d1783c68b5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.953035] env[61867]: DEBUG oslo_concurrency.lockutils [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] Acquired lock "refresh_cache-705a32b6-67f9-42cc-b4d0-f6d1783c68b5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.953203] env[61867]: DEBUG nova.network.neutron [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Refreshing network info cache for port 254a9643-f941-493e-8c87-2932a6cc00ce {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 901.033655] env[61867]: DEBUG oslo_vmware.api [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276793, 'name': ReconfigVM_Task, 'duration_secs': 0.237188} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.033972] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Reconfigured VM instance instance-00000045 to detach disk 2002 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 901.038826] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ef217e8-24a3-4f1f-a6ba-7c05dbdfe7ac {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.054511] env[61867]: DEBUG oslo_vmware.api [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 901.054511] env[61867]: value = "task-1276797" [ 901.054511] env[61867]: _type = "Task" [ 901.054511] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.063405] env[61867]: DEBUG oslo_vmware.api [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276797, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.126560] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.335s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.127289] env[61867]: DEBUG nova.compute.manager [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 901.132683] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.664s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.134220] env[61867]: INFO nova.compute.claims [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 901.154001] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276794, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.274529] env[61867]: DEBUG nova.network.neutron [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Updating instance_info_cache with network_info: [{"id": "af89c243-c119-47c9-9eda-60c9be2b8c51", "address": "fa:16:3e:4d:1f:76", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf89c243-c1", "ovs_interfaceid": "af89c243-c119-47c9-9eda-60c9be2b8c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "65189f98-8c73-47e1-94cd-60c94c3a65d7", "address": "fa:16:3e:f2:05:68", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65189f98-8c", "ovs_interfaceid": "65189f98-8c73-47e1-94cd-60c94c3a65d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a3bb848c-f036-4379-9e73-e7089972da31", "address": "fa:16:3e:6a:34:45", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3bb848c-f0", "ovs_interfaceid": "a3bb848c-f036-4379-9e73-e7089972da31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.282177] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276795, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.312452] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276792, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.322076] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276791, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.409171] env[61867]: DEBUG oslo_vmware.api [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276796, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.564405] env[61867]: DEBUG oslo_vmware.api [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276797, 'name': ReconfigVM_Task, 'duration_secs': 0.148256} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.564753] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274373', 'volume_id': '14e9fde5-2e48-4463-8391-cf95a9fa600a', 'name': 'volume-14e9fde5-2e48-4463-8391-cf95a9fa600a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6dda389d-a8c5-4e0e-87a5-4065e24c034e', 'attached_at': '', 'detached_at': '', 'volume_id': '14e9fde5-2e48-4463-8391-cf95a9fa600a', 'serial': '14e9fde5-2e48-4463-8391-cf95a9fa600a'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 901.639013] env[61867]: DEBUG nova.compute.utils [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 901.642497] env[61867]: DEBUG nova.compute.manager [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 901.642700] env[61867]: DEBUG nova.network.neutron [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 901.654909] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276794, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.705941] env[61867]: DEBUG nova.policy [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ab955c7bc494978bb235ef48300b5ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1dca1c24993044e499ef3ded17361258', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 901.708103] env[61867]: DEBUG nova.network.neutron [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Updated VIF entry in instance network info cache for port 254a9643-f941-493e-8c87-2932a6cc00ce. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 901.708463] env[61867]: DEBUG nova.network.neutron [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Updating instance_info_cache with network_info: [{"id": "254a9643-f941-493e-8c87-2932a6cc00ce", "address": "fa:16:3e:85:01:b4", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap254a9643-f9", "ovs_interfaceid": "254a9643-f941-493e-8c87-2932a6cc00ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.780175] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.780867] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.781087] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.781433] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276795, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.782258] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90be312c-2ab7-4600-a0a8-e00819e4cb2b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.799855] env[61867]: DEBUG nova.virt.hardware [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.800120] env[61867]: DEBUG nova.virt.hardware [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.800349] env[61867]: DEBUG nova.virt.hardware [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.800517] env[61867]: DEBUG nova.virt.hardware [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.800671] env[61867]: DEBUG nova.virt.hardware [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.800826] env[61867]: DEBUG nova.virt.hardware [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.801042] env[61867]: DEBUG nova.virt.hardware [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.801210] env[61867]: DEBUG nova.virt.hardware [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.801406] env[61867]: DEBUG nova.virt.hardware [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.801582] env[61867]: DEBUG nova.virt.hardware [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.801768] env[61867]: DEBUG nova.virt.hardware [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.808012] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Reconfiguring VM to attach interface {{(pid=61867) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 901.808400] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac0843cc-a38a-412c-b5e9-1e8411beef56 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.831710] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276792, 'name': CreateVM_Task, 'duration_secs': 1.106173} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.835293] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.835581] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276791, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.835858] env[61867]: DEBUG oslo_vmware.api [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 901.835858] env[61867]: value = "task-1276798" [ 901.835858] env[61867]: _type = "Task" [ 901.835858] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.836473] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.836637] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.836946] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 901.837469] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37a51d19-5562-48ca-81bd-3c4a0fb2ce76 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.847347] env[61867]: DEBUG oslo_vmware.api [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276798, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.847631] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 901.847631] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ee443b-06bd-72cd-f181-c5b0a394fb61" [ 901.847631] env[61867]: _type = "Task" [ 901.847631] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.854912] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ee443b-06bd-72cd-f181-c5b0a394fb61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.907635] env[61867]: DEBUG oslo_vmware.api [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276796, 'name': PowerOffVM_Task, 'duration_secs': 0.641278} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.907953] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 901.908146] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 901.908406] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf82d0ce-782c-4785-892b-1c5176193818 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.951941] env[61867]: DEBUG nova.network.neutron [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Successfully created port: db6e6b48-df38-493c-bc63-881519b16b64 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 901.962747] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 901.962918] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 901.963130] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Deleting the datastore file [datastore2] b8e561df-069e-4873-91c3-1bfe5f27de2d {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 901.963401] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-973ff873-76c0-417b-a556-26c7fbd742a6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.970736] env[61867]: DEBUG oslo_vmware.api [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for the task: (returnval){ [ 901.970736] env[61867]: value = "task-1276800" [ 901.970736] env[61867]: _type = "Task" [ 901.970736] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.979356] env[61867]: DEBUG oslo_vmware.api [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276800, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.116272] env[61867]: DEBUG nova.objects.instance [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lazy-loading 'flavor' on Instance uuid 6dda389d-a8c5-4e0e-87a5-4065e24c034e {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.148570] env[61867]: DEBUG nova.compute.manager [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 902.162412] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276794, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.211644] env[61867]: DEBUG oslo_concurrency.lockutils [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] Releasing lock "refresh_cache-705a32b6-67f9-42cc-b4d0-f6d1783c68b5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.212471] env[61867]: DEBUG nova.compute.manager [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Received event network-vif-plugged-a3bb848c-f036-4379-9e73-e7089972da31 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.212471] env[61867]: DEBUG oslo_concurrency.lockutils [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] Acquiring lock "ca2f6d18-f773-4875-83de-2f2be912c2f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.212471] env[61867]: DEBUG oslo_concurrency.lockutils [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] Lock "ca2f6d18-f773-4875-83de-2f2be912c2f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.212816] env[61867]: DEBUG oslo_concurrency.lockutils [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] Lock "ca2f6d18-f773-4875-83de-2f2be912c2f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.212816] env[61867]: DEBUG nova.compute.manager [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] No waiting events found dispatching network-vif-plugged-a3bb848c-f036-4379-9e73-e7089972da31 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 902.213053] env[61867]: WARNING nova.compute.manager [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Received unexpected event network-vif-plugged-a3bb848c-f036-4379-9e73-e7089972da31 for instance with vm_state active and task_state None. [ 902.213284] env[61867]: DEBUG nova.compute.manager [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Received event network-changed-a3bb848c-f036-4379-9e73-e7089972da31 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.213521] env[61867]: DEBUG nova.compute.manager [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Refreshing instance network info cache due to event network-changed-a3bb848c-f036-4379-9e73-e7089972da31. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 902.213767] env[61867]: DEBUG oslo_concurrency.lockutils [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] Acquiring lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.213915] env[61867]: DEBUG oslo_concurrency.lockutils [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] Acquired lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.214094] env[61867]: DEBUG nova.network.neutron [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Refreshing network info cache for port a3bb848c-f036-4379-9e73-e7089972da31 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 902.283035] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276795, 'name': CreateSnapshot_Task, 'duration_secs': 1.080342} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.283427] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Created Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 902.284127] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6693258-dbf2-4492-a103-0854945d3615 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.334332] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276791, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.1691} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.338529] env[61867]: INFO nova.virt.vmwareapi.ds_util [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_a6ec9839-2551-4776-a3fb-5540d4202dab/OSTACK_IMG_a6ec9839-2551-4776-a3fb-5540d4202dab.vmdk to [datastore1] devstack-image-cache_base/444eb1e0-4f45-40cf-b139-01bd39317edd/444eb1e0-4f45-40cf-b139-01bd39317edd.vmdk. [ 902.338814] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Cleaning up location [datastore1] OSTACK_IMG_a6ec9839-2551-4776-a3fb-5540d4202dab {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 902.339035] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_a6ec9839-2551-4776-a3fb-5540d4202dab {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 902.339924] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d441fe1b-5dca-4141-ae11-b36f37bd53a8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.358673] env[61867]: DEBUG oslo_vmware.api [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276798, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.359075] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 902.359075] env[61867]: value = "task-1276801" [ 902.359075] env[61867]: _type = "Task" [ 902.359075] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.366315] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ee443b-06bd-72cd-f181-c5b0a394fb61, 'name': SearchDatastore_Task, 'duration_secs': 0.04813} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.369153] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.369482] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.369781] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.369968] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.370227] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.373688] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4ae4d4d7-98d0-4186-8056-512703128642 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.375862] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276801, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.394195] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.394408] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 902.396339] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfb29fd7-9623-4a3e-940a-13b7d55dfb4e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.399692] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f62b77a6-f6ca-449a-aa13-ff3156393343 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.405510] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 902.405510] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525adc59-bb6d-ec3d-9db7-487874d47183" [ 902.405510] env[61867]: _type = "Task" [ 902.405510] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.412551] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a601ae8d-765d-48bb-9b55-23331e78f981 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.421901] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525adc59-bb6d-ec3d-9db7-487874d47183, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.447031] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b6e5c2-5854-4963-973f-7bd2e9a506db {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.454501] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-679d3c29-623f-44ea-8ee1-02e7928b4138 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.469145] env[61867]: DEBUG nova.compute.provider_tree [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.480967] env[61867]: DEBUG oslo_vmware.api [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276800, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.665562] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276794, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.78765} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.666054] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 9e1cbfa9-28e9-4bca-adfd-78bf25428106/9e1cbfa9-28e9-4bca-adfd-78bf25428106.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 902.666213] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 902.666360] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-daced5df-e943-4008-b3f9-b7499420ff7a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.673028] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 902.673028] env[61867]: value = "task-1276802" [ 902.673028] env[61867]: _type = "Task" [ 902.673028] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.681542] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276802, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.807192] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Creating linked-clone VM from snapshot {{(pid=61867) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 902.807567] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-290f91f8-5656-493f-980c-61f9e5e9b15c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.816135] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 902.816135] env[61867]: value = "task-1276803" [ 902.816135] env[61867]: _type = "Task" [ 902.816135] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.825991] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276803, 'name': CloneVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.848761] env[61867]: DEBUG oslo_vmware.api [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276798, 'name': ReconfigVM_Task, 'duration_secs': 0.650954} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.849269] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.849496] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Reconfigured VM to attach interface {{(pid=61867) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 902.870610] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276801, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146586} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.871585] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 902.871766] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Releasing lock "[datastore1] devstack-image-cache_base/444eb1e0-4f45-40cf-b139-01bd39317edd/444eb1e0-4f45-40cf-b139-01bd39317edd.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.872041] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/444eb1e0-4f45-40cf-b139-01bd39317edd/444eb1e0-4f45-40cf-b139-01bd39317edd.vmdk to [datastore1] 1df8427c-e75d-4b60-a92a-b5ba76b67081/1df8427c-e75d-4b60-a92a-b5ba76b67081.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 902.872366] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40976da6-8d03-42bd-83e3-19cbd3c5384b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.879359] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 902.879359] env[61867]: value = "task-1276804" [ 902.879359] env[61867]: _type = "Task" [ 902.879359] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.887519] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276804, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.919736] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525adc59-bb6d-ec3d-9db7-487874d47183, 'name': SearchDatastore_Task, 'duration_secs': 0.031831} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.920687] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b51361c8-2583-4eb1-9bab-56ae08165509 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.926267] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 902.926267] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522ff550-a187-df53-0252-5a666bcb4688" [ 902.926267] env[61867]: _type = "Task" [ 902.926267] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.934186] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522ff550-a187-df53-0252-5a666bcb4688, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.935581] env[61867]: DEBUG nova.network.neutron [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Updated VIF entry in instance network info cache for port a3bb848c-f036-4379-9e73-e7089972da31. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 902.936036] env[61867]: DEBUG nova.network.neutron [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Updating instance_info_cache with network_info: [{"id": "af89c243-c119-47c9-9eda-60c9be2b8c51", "address": "fa:16:3e:4d:1f:76", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf89c243-c1", "ovs_interfaceid": "af89c243-c119-47c9-9eda-60c9be2b8c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "65189f98-8c73-47e1-94cd-60c94c3a65d7", "address": "fa:16:3e:f2:05:68", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65189f98-8c", "ovs_interfaceid": "65189f98-8c73-47e1-94cd-60c94c3a65d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a3bb848c-f036-4379-9e73-e7089972da31", "address": "fa:16:3e:6a:34:45", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3bb848c-f0", "ovs_interfaceid": "a3bb848c-f036-4379-9e73-e7089972da31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.973130] env[61867]: DEBUG nova.scheduler.client.report [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 902.987263] env[61867]: DEBUG oslo_vmware.api [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Task: {'id': task-1276800, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.737453} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.988122] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 902.988323] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 902.988539] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 902.988730] env[61867]: INFO nova.compute.manager [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Took 2.11 seconds to destroy the instance on the hypervisor. [ 902.988968] env[61867]: DEBUG oslo.service.loopingcall [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 902.989403] env[61867]: DEBUG nova.compute.manager [-] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 902.989502] env[61867]: DEBUG nova.network.neutron [-] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 903.123873] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f99af7de-e327-4f4c-9995-1f97bc4ceb28 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.264s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.161483] env[61867]: DEBUG nova.compute.manager [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 903.188333] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276802, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062477} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.190561] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 903.191633] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac052c3-94c0-476e-83c1-c8458d5ca3cf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.197579] env[61867]: DEBUG nova.virt.hardware [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 903.197824] env[61867]: DEBUG nova.virt.hardware [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 903.197985] env[61867]: DEBUG nova.virt.hardware [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 903.198193] env[61867]: DEBUG nova.virt.hardware [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 903.198348] env[61867]: DEBUG nova.virt.hardware [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 903.198501] env[61867]: DEBUG nova.virt.hardware [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 903.198723] env[61867]: DEBUG nova.virt.hardware [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 903.198887] env[61867]: DEBUG nova.virt.hardware [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 903.199070] env[61867]: DEBUG nova.virt.hardware [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 903.199259] env[61867]: DEBUG nova.virt.hardware [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 903.199472] env[61867]: DEBUG nova.virt.hardware [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 903.200351] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d261785-ef47-4ee3-b092-691e6adc8e72 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.220882] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 9e1cbfa9-28e9-4bca-adfd-78bf25428106/9e1cbfa9-28e9-4bca-adfd-78bf25428106.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 903.221692] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e19da30-71a4-460c-97af-cadb1923117f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.240338] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d276b4ca-305c-4c58-a2e5-a574ed3c1416 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.245522] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 903.245522] env[61867]: value = "task-1276805" [ 903.245522] env[61867]: _type = "Task" [ 903.245522] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.261982] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276805, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.328818] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276803, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.331194] env[61867]: DEBUG nova.compute.manager [req-82969833-277e-43e7-a71b-e80fa162149e req-ee7dd535-be82-4aea-93b1-c7e35b364217 service nova] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Received event network-vif-deleted-41bb6052-bacd-4d6b-abd6-b5186a2418e7 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.331426] env[61867]: INFO nova.compute.manager [req-82969833-277e-43e7-a71b-e80fa162149e req-ee7dd535-be82-4aea-93b1-c7e35b364217 service nova] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Neutron deleted interface 41bb6052-bacd-4d6b-abd6-b5186a2418e7; detaching it from the instance and deleting it from the info cache [ 903.331636] env[61867]: DEBUG nova.network.neutron [req-82969833-277e-43e7-a71b-e80fa162149e req-ee7dd535-be82-4aea-93b1-c7e35b364217 service nova] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.354134] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9863d01a-c169-4dcc-8f95-6755067e5f27 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "interface-ca2f6d18-f773-4875-83de-2f2be912c2f8-a3bb848c-f036-4379-9e73-e7089972da31" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 5.939s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.389563] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276804, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.438460] env[61867]: DEBUG oslo_concurrency.lockutils [req-519f6f78-49df-4878-b449-3c2f69e54869 req-ea4c19ab-d977-4cca-aa6a-ec5b3a4aa226 service nova] Releasing lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.438945] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522ff550-a187-df53-0252-5a666bcb4688, 'name': SearchDatastore_Task, 'duration_secs': 0.024971} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.439279] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.439575] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 705a32b6-67f9-42cc-b4d0-f6d1783c68b5/705a32b6-67f9-42cc-b4d0-f6d1783c68b5.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 903.439872] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-de515962-e2da-452c-8c25-5cb47df9e748 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.450441] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 903.450441] env[61867]: value = "task-1276806" [ 903.450441] env[61867]: _type = "Task" [ 903.450441] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.462741] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276806, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.483397] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.351s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.484146] env[61867]: DEBUG nova.compute.manager [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 903.508466] env[61867]: DEBUG nova.network.neutron [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Successfully updated port: db6e6b48-df38-493c-bc63-881519b16b64 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 903.758709] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276805, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.771444] env[61867]: DEBUG nova.network.neutron [-] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.829527] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276803, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.834500] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-62022ec3-035f-4b04-9787-482abc28e34a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.845987] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b39e09de-5ee6-45e2-84a3-4b3c7ba942b8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.882033] env[61867]: DEBUG nova.compute.manager [req-82969833-277e-43e7-a71b-e80fa162149e req-ee7dd535-be82-4aea-93b1-c7e35b364217 service nova] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Detach interface failed, port_id=41bb6052-bacd-4d6b-abd6-b5186a2418e7, reason: Instance b8e561df-069e-4873-91c3-1bfe5f27de2d could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 903.892069] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276804, 'name': CopyVirtualDisk_Task} progress is 26%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.962421] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276806, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.989241] env[61867]: DEBUG nova.compute.utils [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 903.991063] env[61867]: DEBUG nova.compute.manager [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 903.991241] env[61867]: DEBUG nova.network.neutron [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 904.011634] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.011634] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquired lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.011811] env[61867]: DEBUG nova.network.neutron [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 904.041948] env[61867]: DEBUG nova.policy [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'adcd9eb75ecc4eccb335ebfec207a900', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7316e4f263a9432ab2f9f91484d62f58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 904.218421] env[61867]: DEBUG oslo_concurrency.lockutils [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "6d2dab88-4165-4952-8019-2eaf3b863115" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.218606] env[61867]: DEBUG oslo_concurrency.lockutils [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "6d2dab88-4165-4952-8019-2eaf3b863115" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.259474] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276805, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.273832] env[61867]: INFO nova.compute.manager [-] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Took 1.28 seconds to deallocate network for instance. [ 904.296735] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.297080] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.297385] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.297654] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.297860] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.303021] env[61867]: INFO nova.compute.manager [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Terminating instance [ 904.304361] env[61867]: DEBUG nova.compute.manager [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 904.304579] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 904.305594] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-124d7dfb-7701-44fc-96cd-07097f244a58 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.316095] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 904.316095] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e9096ad-2be4-42cf-bd46-65e3e37216c8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.324865] env[61867]: DEBUG oslo_vmware.api [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 904.324865] env[61867]: value = "task-1276807" [ 904.324865] env[61867]: _type = "Task" [ 904.324865] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.332477] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276803, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.342798] env[61867]: DEBUG oslo_vmware.api [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276807, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.349540] env[61867]: DEBUG nova.network.neutron [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Successfully created port: f91d6c97-5188-4d51-8b56-411f6ca824c2 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 904.394018] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276804, 'name': CopyVirtualDisk_Task} progress is 49%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.462123] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276806, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.497039] env[61867]: DEBUG nova.compute.manager [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 904.581219] env[61867]: DEBUG nova.network.neutron [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 904.722290] env[61867]: DEBUG nova.compute.utils [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 904.757624] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276805, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.781737] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.782141] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.782257] env[61867]: DEBUG nova.objects.instance [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lazy-loading 'resources' on Instance uuid b8e561df-069e-4873-91c3-1bfe5f27de2d {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.784262] env[61867]: DEBUG nova.network.neutron [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Updating instance_info_cache with network_info: [{"id": "db6e6b48-df38-493c-bc63-881519b16b64", "address": "fa:16:3e:da:a7:2a", "network": {"id": "d8a081a0-0d77-4ae7-9d58-19da456f7699", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-201249363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dca1c24993044e499ef3ded17361258", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb6e6b48-df", "ovs_interfaceid": "db6e6b48-df38-493c-bc63-881519b16b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.832908] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276803, 'name': CloneVM_Task, 'duration_secs': 1.947661} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.833700] env[61867]: INFO nova.virt.vmwareapi.vmops [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Created linked-clone VM from snapshot [ 904.834554] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf8df94-aac4-42c0-9525-8e4cd4183a90 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.841164] env[61867]: DEBUG oslo_vmware.api [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276807, 'name': PowerOffVM_Task, 'duration_secs': 0.318716} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.844571] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 904.844800] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 904.845353] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Uploading image e21e0ddf-319b-457b-b08e-49317c63fa31 {{(pid=61867) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 904.847369] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ec40a4bf-7c24-42dc-a557-a49cdb4b94de {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.873546] env[61867]: DEBUG oslo_vmware.rw_handles [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 904.873546] env[61867]: value = "vm-274382" [ 904.873546] env[61867]: _type = "VirtualMachine" [ 904.873546] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 904.873878] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-6ae1a567-3f0b-407d-9970-d3d197350605 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.883919] env[61867]: DEBUG oslo_vmware.rw_handles [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lease: (returnval){ [ 904.883919] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52aa3cd1-cacf-b073-9f81-3c20d16bfb79" [ 904.883919] env[61867]: _type = "HttpNfcLease" [ 904.883919] env[61867]: } obtained for exporting VM: (result){ [ 904.883919] env[61867]: value = "vm-274382" [ 904.883919] env[61867]: _type = "VirtualMachine" [ 904.883919] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 904.884354] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the lease: (returnval){ [ 904.884354] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52aa3cd1-cacf-b073-9f81-3c20d16bfb79" [ 904.884354] env[61867]: _type = "HttpNfcLease" [ 904.884354] env[61867]: } to be ready. {{(pid=61867) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 904.898046] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 904.898046] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52aa3cd1-cacf-b073-9f81-3c20d16bfb79" [ 904.898046] env[61867]: _type = "HttpNfcLease" [ 904.898046] env[61867]: } is initializing. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 904.900934] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276804, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.966016] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276806, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.970528] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 904.970765] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 904.970957] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Deleting the datastore file [datastore2] 6dda389d-a8c5-4e0e-87a5-4065e24c034e {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 904.971255] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b27ea7be-2476-4d8a-b5a0-29a435a56fd5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.978640] env[61867]: DEBUG oslo_vmware.api [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for the task: (returnval){ [ 904.978640] env[61867]: value = "task-1276810" [ 904.978640] env[61867]: _type = "Task" [ 904.978640] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.989983] env[61867]: DEBUG oslo_vmware.api [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276810, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.027989] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.028300] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.226058] env[61867]: DEBUG oslo_concurrency.lockutils [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "6d2dab88-4165-4952-8019-2eaf3b863115" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.259083] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276805, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.289599] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Releasing lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.289927] env[61867]: DEBUG nova.compute.manager [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Instance network_info: |[{"id": "db6e6b48-df38-493c-bc63-881519b16b64", "address": "fa:16:3e:da:a7:2a", "network": {"id": "d8a081a0-0d77-4ae7-9d58-19da456f7699", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-201249363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dca1c24993044e499ef3ded17361258", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb6e6b48-df", "ovs_interfaceid": "db6e6b48-df38-493c-bc63-881519b16b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 905.290617] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:a7:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'db6e6b48-df38-493c-bc63-881519b16b64', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.298217] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Creating folder: Project (1dca1c24993044e499ef3ded17361258). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 905.298896] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ad7fe866-ca37-4e5d-9ef5-3cbdffd1ae71 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.313405] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Created folder: Project (1dca1c24993044e499ef3ded17361258) in parent group-v274258. [ 905.314576] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Creating folder: Instances. Parent ref: group-v274383. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 905.314576] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bb553f06-746f-4c03-9e97-b0d102d80879 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.326079] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Created folder: Instances in parent group-v274383. [ 905.326600] env[61867]: DEBUG oslo.service.loopingcall [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.326600] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 905.326993] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ac9e4370-0bb3-47dd-a5df-bc078fb03968 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.351357] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.351357] env[61867]: value = "task-1276813" [ 905.351357] env[61867]: _type = "Task" [ 905.351357] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.364085] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276813, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.397194] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276804, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.404423] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 905.404423] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52aa3cd1-cacf-b073-9f81-3c20d16bfb79" [ 905.404423] env[61867]: _type = "HttpNfcLease" [ 905.404423] env[61867]: } is ready. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 905.404775] env[61867]: DEBUG oslo_vmware.rw_handles [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 905.404775] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52aa3cd1-cacf-b073-9f81-3c20d16bfb79" [ 905.404775] env[61867]: _type = "HttpNfcLease" [ 905.404775] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 905.405560] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6842841-9e92-4f85-9a40-29de646a393b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.419627] env[61867]: DEBUG oslo_vmware.rw_handles [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b55588-fd0c-3a5b-7ac3-14d186a85ed3/disk-0.vmdk from lease info. {{(pid=61867) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 905.419887] env[61867]: DEBUG oslo_vmware.rw_handles [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b55588-fd0c-3a5b-7ac3-14d186a85ed3/disk-0.vmdk for reading. {{(pid=61867) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 905.497454] env[61867]: DEBUG oslo_vmware.api [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Task: {'id': task-1276810, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.274503} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.501154] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 905.501538] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 905.501746] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 905.501942] env[61867]: INFO nova.compute.manager [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Took 1.20 seconds to destroy the instance on the hypervisor. [ 905.502237] env[61867]: DEBUG oslo.service.loopingcall [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.502504] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276806, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.505619] env[61867]: DEBUG nova.compute.manager [-] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 905.505729] env[61867]: DEBUG nova.network.neutron [-] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 905.508372] env[61867]: DEBUG nova.compute.manager [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 905.532841] env[61867]: DEBUG nova.compute.utils [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 905.537521] env[61867]: DEBUG nova.virt.hardware [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 905.537887] env[61867]: DEBUG nova.virt.hardware [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 905.538091] env[61867]: DEBUG nova.virt.hardware [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 905.538372] env[61867]: DEBUG nova.virt.hardware [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 905.538569] env[61867]: DEBUG nova.virt.hardware [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 905.538801] env[61867]: DEBUG nova.virt.hardware [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 905.538950] env[61867]: DEBUG nova.virt.hardware [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 905.539133] env[61867]: DEBUG nova.virt.hardware [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 905.539354] env[61867]: DEBUG nova.virt.hardware [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 905.539591] env[61867]: DEBUG nova.virt.hardware [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 905.539814] env[61867]: DEBUG nova.virt.hardware [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 905.541457] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fd675b33-31bb-4166-a694-fcf9977c2760 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.543757] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b43580d9-e8b8-441d-992b-889d51020295 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.561993] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c13a2e5-db2f-48fc-a047-9efdd8016590 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.635315] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-527eca1a-8918-4d6e-83cf-595d612c544b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.648292] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6dd2d60-d513-487f-be3a-72c9a8b31d4e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.684107] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed88ebff-2b07-4cf0-975c-2628cb62143b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.693820] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-281b6989-7069-4984-a55c-7346cf2c17bc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.713283] env[61867]: DEBUG nova.compute.provider_tree [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.762187] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276805, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.865026] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276813, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.897113] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276804, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.91766} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.898028] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/444eb1e0-4f45-40cf-b139-01bd39317edd/444eb1e0-4f45-40cf-b139-01bd39317edd.vmdk to [datastore1] 1df8427c-e75d-4b60-a92a-b5ba76b67081/1df8427c-e75d-4b60-a92a-b5ba76b67081.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 905.899270] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fc46a90-3cc1-4091-8ab5-dba9c135c869 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.944674] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 1df8427c-e75d-4b60-a92a-b5ba76b67081/1df8427c-e75d-4b60-a92a-b5ba76b67081.vmdk or device None with type streamOptimized {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.945445] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d2a7392-33f0-491c-be9a-bdfbc3c0461b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.971225] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 905.971225] env[61867]: value = "task-1276814" [ 905.971225] env[61867]: _type = "Task" [ 905.971225] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.982713] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276814, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.991684] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276806, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.424838} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.992073] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 705a32b6-67f9-42cc-b4d0-f6d1783c68b5/705a32b6-67f9-42cc-b4d0-f6d1783c68b5.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 905.992340] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 905.992775] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d0e45812-1add-4e26-b6dc-0b22dc63af07 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.003511] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 906.003511] env[61867]: value = "task-1276815" [ 906.003511] env[61867]: _type = "Task" [ 906.003511] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.014849] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276815, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.048221] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.020s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.217182] env[61867]: DEBUG nova.scheduler.client.report [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 906.262185] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276805, 'name': ReconfigVM_Task, 'duration_secs': 2.675507} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.262185] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 9e1cbfa9-28e9-4bca-adfd-78bf25428106/9e1cbfa9-28e9-4bca-adfd-78bf25428106.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.262185] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3588b57f-26ab-45c0-928b-7017b7d8a974 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.273029] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 906.273029] env[61867]: value = "task-1276816" [ 906.273029] env[61867]: _type = "Task" [ 906.273029] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.277829] env[61867]: DEBUG nova.network.neutron [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Successfully updated port: f91d6c97-5188-4d51-8b56-411f6ca824c2 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 906.288759] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276816, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.309971] env[61867]: DEBUG oslo_concurrency.lockutils [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "6d2dab88-4165-4952-8019-2eaf3b863115" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.310277] env[61867]: DEBUG oslo_concurrency.lockutils [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "6d2dab88-4165-4952-8019-2eaf3b863115" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.310547] env[61867]: INFO nova.compute.manager [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Attaching volume 87de05dc-cb88-42e1-a436-004b68dd4e5c to /dev/sdb [ 906.348262] env[61867]: DEBUG nova.compute.manager [req-fcf0ff85-64a1-4260-aeb8-e667190fd886 req-0efe293a-ce3f-4a20-b235-b4f915d54d28 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Received event network-vif-plugged-db6e6b48-df38-493c-bc63-881519b16b64 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.348262] env[61867]: DEBUG oslo_concurrency.lockutils [req-fcf0ff85-64a1-4260-aeb8-e667190fd886 req-0efe293a-ce3f-4a20-b235-b4f915d54d28 service nova] Acquiring lock "816ca796-d8ed-4843-9b1f-f169f48ff047-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.348262] env[61867]: DEBUG oslo_concurrency.lockutils [req-fcf0ff85-64a1-4260-aeb8-e667190fd886 req-0efe293a-ce3f-4a20-b235-b4f915d54d28 service nova] Lock "816ca796-d8ed-4843-9b1f-f169f48ff047-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.348262] env[61867]: DEBUG oslo_concurrency.lockutils [req-fcf0ff85-64a1-4260-aeb8-e667190fd886 req-0efe293a-ce3f-4a20-b235-b4f915d54d28 service nova] Lock "816ca796-d8ed-4843-9b1f-f169f48ff047-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.348262] env[61867]: DEBUG nova.compute.manager [req-fcf0ff85-64a1-4260-aeb8-e667190fd886 req-0efe293a-ce3f-4a20-b235-b4f915d54d28 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] No waiting events found dispatching network-vif-plugged-db6e6b48-df38-493c-bc63-881519b16b64 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 906.348262] env[61867]: WARNING nova.compute.manager [req-fcf0ff85-64a1-4260-aeb8-e667190fd886 req-0efe293a-ce3f-4a20-b235-b4f915d54d28 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Received unexpected event network-vif-plugged-db6e6b48-df38-493c-bc63-881519b16b64 for instance with vm_state building and task_state spawning. [ 906.348964] env[61867]: DEBUG nova.compute.manager [req-fcf0ff85-64a1-4260-aeb8-e667190fd886 req-0efe293a-ce3f-4a20-b235-b4f915d54d28 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Received event network-changed-db6e6b48-df38-493c-bc63-881519b16b64 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.348964] env[61867]: DEBUG nova.compute.manager [req-fcf0ff85-64a1-4260-aeb8-e667190fd886 req-0efe293a-ce3f-4a20-b235-b4f915d54d28 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Refreshing instance network info cache due to event network-changed-db6e6b48-df38-493c-bc63-881519b16b64. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 906.348964] env[61867]: DEBUG oslo_concurrency.lockutils [req-fcf0ff85-64a1-4260-aeb8-e667190fd886 req-0efe293a-ce3f-4a20-b235-b4f915d54d28 service nova] Acquiring lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.349078] env[61867]: DEBUG oslo_concurrency.lockutils [req-fcf0ff85-64a1-4260-aeb8-e667190fd886 req-0efe293a-ce3f-4a20-b235-b4f915d54d28 service nova] Acquired lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.349367] env[61867]: DEBUG nova.network.neutron [req-fcf0ff85-64a1-4260-aeb8-e667190fd886 req-0efe293a-ce3f-4a20-b235-b4f915d54d28 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Refreshing network info cache for port db6e6b48-df38-493c-bc63-881519b16b64 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 906.361014] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5216dfd7-2676-42a3-bce8-9ff8db8bf2dc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.368772] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276813, 'name': CreateVM_Task, 'duration_secs': 0.648039} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.372872] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 906.372872] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.373134] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.373596] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 906.374425] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14ef4a6d-1720-41ab-a03a-3843247cfb12 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.382200] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-565eb225-954f-4662-8e47-75d2a3d8d0dd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.389750] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 906.389750] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528e261d-7c26-360e-f548-5c28baa6ccec" [ 906.389750] env[61867]: _type = "Task" [ 906.389750] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.403723] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528e261d-7c26-360e-f548-5c28baa6ccec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.411422] env[61867]: DEBUG nova.virt.block_device [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Updating existing volume attachment record: bc147517-92a1-4f79-8e68-a0552aee98c6 {{(pid=61867) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 906.484112] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276814, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.522032] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276815, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.197194} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.522032] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 906.522032] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda99a82-d6d5-479c-ac32-d1f5537ceb44 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.544976] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 705a32b6-67f9-42cc-b4d0-f6d1783c68b5/705a32b6-67f9-42cc-b4d0-f6d1783c68b5.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 906.545468] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bf571648-c8e5-4440-8ab5-29e4aba5992c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.568737] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 906.568737] env[61867]: value = "task-1276817" [ 906.568737] env[61867]: _type = "Task" [ 906.568737] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.579420] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276817, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.723410] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.941s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.760805] env[61867]: INFO nova.scheduler.client.report [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Deleted allocations for instance b8e561df-069e-4873-91c3-1bfe5f27de2d [ 906.781489] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "refresh_cache-5ec9eb1f-1d24-41d9-b5c7-a478a427c563" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.784045] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "refresh_cache-5ec9eb1f-1d24-41d9-b5c7-a478a427c563" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.784045] env[61867]: DEBUG nova.network.neutron [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 906.792986] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276816, 'name': Rename_Task, 'duration_secs': 0.399315} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.794511] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 906.795395] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73e88994-15a8-4374-ae5f-35eceef34f33 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.806728] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 906.806728] env[61867]: value = "task-1276821" [ 906.806728] env[61867]: _type = "Task" [ 906.806728] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.811337] env[61867]: DEBUG nova.network.neutron [-] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.819296] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276821, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.915261] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528e261d-7c26-360e-f548-5c28baa6ccec, 'name': SearchDatastore_Task, 'duration_secs': 0.031796} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.916026] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.916590] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 906.917534] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.917534] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.918084] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 906.918206] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bf730bc5-4363-4fa1-8a61-dc3094b3bbec {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.931588] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 906.932051] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 906.933251] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a2f23e0-514d-42ca-966b-292d5d127ebd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.942585] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 906.942585] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522ac8eb-48a3-4ddb-5425-34d6a878c09e" [ 906.942585] env[61867]: _type = "Task" [ 906.942585] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.957880] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522ac8eb-48a3-4ddb-5425-34d6a878c09e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.990395] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276814, 'name': ReconfigVM_Task, 'duration_secs': 0.909746} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.990780] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 1df8427c-e75d-4b60-a92a-b5ba76b67081/1df8427c-e75d-4b60-a92a-b5ba76b67081.vmdk or device None with type streamOptimized {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.991791] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2cc288d6-069f-411e-b503-ee1f1a45a6d6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.001964] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 907.001964] env[61867]: value = "task-1276822" [ 907.001964] env[61867]: _type = "Task" [ 907.001964] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.023036] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276822, 'name': Rename_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.080446] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276817, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.119832] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.120273] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.120559] env[61867]: INFO nova.compute.manager [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Attaching volume 5c23a596-074b-4f75-a4bc-86c77ffc5080 to /dev/sdb [ 907.168302] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-432b789a-b1f7-4b2d-a1ed-0fad37aac01f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.178776] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f0a549-32be-49a4-aa70-1825d8d1ffc7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.205667] env[61867]: DEBUG nova.virt.block_device [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Updating existing volume attachment record: 910a3d28-585c-45a7-bb13-1acacc14c91c {{(pid=61867) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 907.269983] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0bd246df-6733-4afb-ad60-434bf4349614 tempest-ImagesTestJSON-1934539458 tempest-ImagesTestJSON-1934539458-project-member] Lock "b8e561df-069e-4873-91c3-1bfe5f27de2d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.395s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.319297] env[61867]: INFO nova.compute.manager [-] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Took 1.81 seconds to deallocate network for instance. [ 907.319297] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276821, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.328061] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "interface-ca2f6d18-f773-4875-83de-2f2be912c2f8-65189f98-8c73-47e1-94cd-60c94c3a65d7" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.328400] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "interface-ca2f6d18-f773-4875-83de-2f2be912c2f8-65189f98-8c73-47e1-94cd-60c94c3a65d7" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.358046] env[61867]: DEBUG nova.network.neutron [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 907.428391] env[61867]: DEBUG nova.network.neutron [req-fcf0ff85-64a1-4260-aeb8-e667190fd886 req-0efe293a-ce3f-4a20-b235-b4f915d54d28 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Updated VIF entry in instance network info cache for port db6e6b48-df38-493c-bc63-881519b16b64. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 907.428845] env[61867]: DEBUG nova.network.neutron [req-fcf0ff85-64a1-4260-aeb8-e667190fd886 req-0efe293a-ce3f-4a20-b235-b4f915d54d28 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Updating instance_info_cache with network_info: [{"id": "db6e6b48-df38-493c-bc63-881519b16b64", "address": "fa:16:3e:da:a7:2a", "network": {"id": "d8a081a0-0d77-4ae7-9d58-19da456f7699", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-201249363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dca1c24993044e499ef3ded17361258", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb6e6b48-df", "ovs_interfaceid": "db6e6b48-df38-493c-bc63-881519b16b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.460544] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522ac8eb-48a3-4ddb-5425-34d6a878c09e, 'name': SearchDatastore_Task, 'duration_secs': 0.032217} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.461268] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-144ebf03-fc21-4077-8496-9e88cbe8cacf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.469208] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 907.469208] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521c378b-6204-3501-3fc5-579db5748330" [ 907.469208] env[61867]: _type = "Task" [ 907.469208] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.481078] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521c378b-6204-3501-3fc5-579db5748330, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.513471] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276822, 'name': Rename_Task, 'duration_secs': 0.222352} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.513818] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 907.514203] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2745798f-ae19-47b0-a6eb-c43a484afc76 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.524486] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 907.524486] env[61867]: value = "task-1276825" [ 907.524486] env[61867]: _type = "Task" [ 907.524486] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.536635] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276825, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.580729] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276817, 'name': ReconfigVM_Task, 'duration_secs': 0.64751} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.581925] env[61867]: DEBUG nova.network.neutron [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Updating instance_info_cache with network_info: [{"id": "f91d6c97-5188-4d51-8b56-411f6ca824c2", "address": "fa:16:3e:fb:30:c8", "network": {"id": "14cee476-296a-4093-99ba-c2a5ef63f7cf", "bridge": "br-int", "label": "tempest-ServersTestJSON-243183479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7316e4f263a9432ab2f9f91484d62f58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf91d6c97-51", "ovs_interfaceid": "f91d6c97-5188-4d51-8b56-411f6ca824c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.583606] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 705a32b6-67f9-42cc-b4d0-f6d1783c68b5/705a32b6-67f9-42cc-b4d0-f6d1783c68b5.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 907.584859] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-91c5b7e8-1510-47ac-9b42-7053ffe8929d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.593928] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 907.593928] env[61867]: value = "task-1276827" [ 907.593928] env[61867]: _type = "Task" [ 907.593928] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.604968] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276827, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.820968] env[61867]: DEBUG oslo_vmware.api [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276821, 'name': PowerOnVM_Task, 'duration_secs': 0.761319} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.821478] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 907.821788] env[61867]: INFO nova.compute.manager [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Took 12.40 seconds to spawn the instance on the hypervisor. [ 907.822117] env[61867]: DEBUG nova.compute.manager [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 907.823894] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.824168] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.824519] env[61867]: DEBUG nova.objects.instance [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lazy-loading 'resources' on Instance uuid 6dda389d-a8c5-4e0e-87a5-4065e24c034e {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.826184] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-792c92c9-85cf-4d47-b18d-3cce0a830b82 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.832147] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.832340] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.834845] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe955c6-10d9-4a33-96ca-0d8fd0f52ab4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.862863] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e9825d-3002-43e8-9b2f-f6dc3b853825 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.896314] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Reconfiguring VM to detach interface {{(pid=61867) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 907.897038] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1880be21-c1d6-47d5-b107-5183e9bdb6f0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.920859] env[61867]: DEBUG oslo_vmware.api [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 907.920859] env[61867]: value = "task-1276828" [ 907.920859] env[61867]: _type = "Task" [ 907.920859] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.931484] env[61867]: DEBUG oslo_concurrency.lockutils [req-fcf0ff85-64a1-4260-aeb8-e667190fd886 req-0efe293a-ce3f-4a20-b235-b4f915d54d28 service nova] Releasing lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.931940] env[61867]: DEBUG oslo_vmware.api [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276828, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.981023] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521c378b-6204-3501-3fc5-579db5748330, 'name': SearchDatastore_Task, 'duration_secs': 0.014352} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.981354] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.981705] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 816ca796-d8ed-4843-9b1f-f169f48ff047/816ca796-d8ed-4843-9b1f-f169f48ff047.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 907.981999] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7691e6ec-fa4f-4eb2-9abb-efd964fc641a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.991307] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 907.991307] env[61867]: value = "task-1276829" [ 907.991307] env[61867]: _type = "Task" [ 907.991307] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.001735] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276829, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.037157] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276825, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.086136] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "refresh_cache-5ec9eb1f-1d24-41d9-b5c7-a478a427c563" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.086136] env[61867]: DEBUG nova.compute.manager [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Instance network_info: |[{"id": "f91d6c97-5188-4d51-8b56-411f6ca824c2", "address": "fa:16:3e:fb:30:c8", "network": {"id": "14cee476-296a-4093-99ba-c2a5ef63f7cf", "bridge": "br-int", "label": "tempest-ServersTestJSON-243183479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7316e4f263a9432ab2f9f91484d62f58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf91d6c97-51", "ovs_interfaceid": "f91d6c97-5188-4d51-8b56-411f6ca824c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 908.086619] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:30:c8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '085fb0ff-9285-4f1d-a008-a14da4844357', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f91d6c97-5188-4d51-8b56-411f6ca824c2', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 908.096632] env[61867]: DEBUG oslo.service.loopingcall [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.096974] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 908.101463] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2cd02983-ee73-4170-ac99-b01012f05301 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.124217] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276827, 'name': Rename_Task, 'duration_secs': 0.194796} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.125514] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 908.125756] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 908.125756] env[61867]: value = "task-1276830" [ 908.125756] env[61867]: _type = "Task" [ 908.125756] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.125977] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd28d770-b312-43fd-bf19-bc6a3fddeb41 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.138866] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276830, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.140510] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 908.140510] env[61867]: value = "task-1276831" [ 908.140510] env[61867]: _type = "Task" [ 908.140510] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.148854] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276831, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.353384] env[61867]: INFO nova.compute.manager [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Took 23.73 seconds to build instance. [ 908.391718] env[61867]: DEBUG nova.compute.manager [req-9fcb3257-c081-4d7e-a389-4e6a4549938b req-1fdd0ffe-8711-4a6d-9d3d-32f30672380e service nova] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Received event network-vif-plugged-f91d6c97-5188-4d51-8b56-411f6ca824c2 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.392144] env[61867]: DEBUG oslo_concurrency.lockutils [req-9fcb3257-c081-4d7e-a389-4e6a4549938b req-1fdd0ffe-8711-4a6d-9d3d-32f30672380e service nova] Acquiring lock "5ec9eb1f-1d24-41d9-b5c7-a478a427c563-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.392603] env[61867]: DEBUG oslo_concurrency.lockutils [req-9fcb3257-c081-4d7e-a389-4e6a4549938b req-1fdd0ffe-8711-4a6d-9d3d-32f30672380e service nova] Lock "5ec9eb1f-1d24-41d9-b5c7-a478a427c563-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.392961] env[61867]: DEBUG oslo_concurrency.lockutils [req-9fcb3257-c081-4d7e-a389-4e6a4549938b req-1fdd0ffe-8711-4a6d-9d3d-32f30672380e service nova] Lock "5ec9eb1f-1d24-41d9-b5c7-a478a427c563-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.393298] env[61867]: DEBUG nova.compute.manager [req-9fcb3257-c081-4d7e-a389-4e6a4549938b req-1fdd0ffe-8711-4a6d-9d3d-32f30672380e service nova] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] No waiting events found dispatching network-vif-plugged-f91d6c97-5188-4d51-8b56-411f6ca824c2 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 908.393612] env[61867]: WARNING nova.compute.manager [req-9fcb3257-c081-4d7e-a389-4e6a4549938b req-1fdd0ffe-8711-4a6d-9d3d-32f30672380e service nova] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Received unexpected event network-vif-plugged-f91d6c97-5188-4d51-8b56-411f6ca824c2 for instance with vm_state building and task_state spawning. [ 908.393954] env[61867]: DEBUG nova.compute.manager [req-9fcb3257-c081-4d7e-a389-4e6a4549938b req-1fdd0ffe-8711-4a6d-9d3d-32f30672380e service nova] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Received event network-changed-f91d6c97-5188-4d51-8b56-411f6ca824c2 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.394258] env[61867]: DEBUG nova.compute.manager [req-9fcb3257-c081-4d7e-a389-4e6a4549938b req-1fdd0ffe-8711-4a6d-9d3d-32f30672380e service nova] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Refreshing instance network info cache due to event network-changed-f91d6c97-5188-4d51-8b56-411f6ca824c2. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 908.394625] env[61867]: DEBUG oslo_concurrency.lockutils [req-9fcb3257-c081-4d7e-a389-4e6a4549938b req-1fdd0ffe-8711-4a6d-9d3d-32f30672380e service nova] Acquiring lock "refresh_cache-5ec9eb1f-1d24-41d9-b5c7-a478a427c563" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.394894] env[61867]: DEBUG oslo_concurrency.lockutils [req-9fcb3257-c081-4d7e-a389-4e6a4549938b req-1fdd0ffe-8711-4a6d-9d3d-32f30672380e service nova] Acquired lock "refresh_cache-5ec9eb1f-1d24-41d9-b5c7-a478a427c563" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.395231] env[61867]: DEBUG nova.network.neutron [req-9fcb3257-c081-4d7e-a389-4e6a4549938b req-1fdd0ffe-8711-4a6d-9d3d-32f30672380e service nova] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Refreshing network info cache for port f91d6c97-5188-4d51-8b56-411f6ca824c2 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 908.434802] env[61867]: DEBUG oslo_vmware.api [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276828, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.508337] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276829, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.536381] env[61867]: DEBUG oslo_vmware.api [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276825, 'name': PowerOnVM_Task, 'duration_secs': 0.585661} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.536736] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 908.600276] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab9056d-4197-4a44-9609-1aba23c0960c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.608063] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f9aa9f1-807d-4514-ae01-73051ae1c820 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.649704] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-717b904f-245b-4083-beb7-f1006ca2a3d7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.664297] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-368cbb81-e2bc-4517-addf-672c79c4af8d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.673081] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276830, 'name': CreateVM_Task, 'duration_secs': 0.46686} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.674663] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276831, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.674663] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 908.675641] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.675813] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.676310] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 908.676437] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-030d7a5e-0a7b-497e-9a75-e7eca348b63e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.687086] env[61867]: DEBUG nova.compute.provider_tree [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.693942] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 908.693942] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52abefb6-dcd8-27f6-d9c6-d7a6d4e31862" [ 908.693942] env[61867]: _type = "Task" [ 908.693942] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.705534] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52abefb6-dcd8-27f6-d9c6-d7a6d4e31862, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.856222] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c33cdba5-97e9-417b-89a6-4a5a633b8fce tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.236s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.910524] env[61867]: DEBUG nova.compute.manager [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 908.911306] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72fbd6bb-665f-4e10-8b44-d5759945ca4f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.932979] env[61867]: DEBUG oslo_vmware.api [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276828, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.008595] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276829, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539832} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.008961] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 816ca796-d8ed-4843-9b1f-f169f48ff047/816ca796-d8ed-4843-9b1f-f169f48ff047.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 909.009288] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 909.009635] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de30fba0-6e30-4d64-866b-18c47b43efa8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.017178] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 909.017178] env[61867]: value = "task-1276833" [ 909.017178] env[61867]: _type = "Task" [ 909.017178] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.028969] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276833, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.154599] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276831, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.191897] env[61867]: DEBUG nova.scheduler.client.report [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.205388] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52abefb6-dcd8-27f6-d9c6-d7a6d4e31862, 'name': SearchDatastore_Task, 'duration_secs': 0.016064} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.205709] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.205940] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 909.206199] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.206351] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.206538] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 909.206808] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99bac4cb-3423-4b43-b7bc-dc91a6c6fce4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.225101] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 909.225415] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 909.227290] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-947811f0-b269-4a3a-b899-6057baed7ac8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.233098] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 909.233098] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f26058-b07a-fe2f-490d-41b5b01b889e" [ 909.233098] env[61867]: _type = "Task" [ 909.233098] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.245021] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f26058-b07a-fe2f-490d-41b5b01b889e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.277676] env[61867]: DEBUG nova.network.neutron [req-9fcb3257-c081-4d7e-a389-4e6a4549938b req-1fdd0ffe-8711-4a6d-9d3d-32f30672380e service nova] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Updated VIF entry in instance network info cache for port f91d6c97-5188-4d51-8b56-411f6ca824c2. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 909.277676] env[61867]: DEBUG nova.network.neutron [req-9fcb3257-c081-4d7e-a389-4e6a4549938b req-1fdd0ffe-8711-4a6d-9d3d-32f30672380e service nova] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Updating instance_info_cache with network_info: [{"id": "f91d6c97-5188-4d51-8b56-411f6ca824c2", "address": "fa:16:3e:fb:30:c8", "network": {"id": "14cee476-296a-4093-99ba-c2a5ef63f7cf", "bridge": "br-int", "label": "tempest-ServersTestJSON-243183479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7316e4f263a9432ab2f9f91484d62f58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf91d6c97-51", "ovs_interfaceid": "f91d6c97-5188-4d51-8b56-411f6ca824c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.321951] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.321951] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.321951] env[61867]: INFO nova.compute.manager [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Shelving [ 909.433130] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0009c3bb-1224-4a55-8f17-78014ca4639b tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "1df8427c-e75d-4b60-a92a-b5ba76b67081" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 32.369s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.439590] env[61867]: DEBUG oslo_vmware.api [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276828, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.529797] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276833, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080363} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.530108] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 909.531225] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffbd3d89-b06b-48d2-af37-bb2147f24ad8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.555382] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 816ca796-d8ed-4843-9b1f-f169f48ff047/816ca796-d8ed-4843-9b1f-f169f48ff047.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 909.555737] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ecc97153-d913-422f-90e5-df82fa3d0eab {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.580854] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 909.580854] env[61867]: value = "task-1276834" [ 909.580854] env[61867]: _type = "Task" [ 909.580854] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.591692] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276834, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.656793] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276831, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.700150] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.876s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.727595] env[61867]: INFO nova.scheduler.client.report [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Deleted allocations for instance 6dda389d-a8c5-4e0e-87a5-4065e24c034e [ 909.743999] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f26058-b07a-fe2f-490d-41b5b01b889e, 'name': SearchDatastore_Task, 'duration_secs': 0.021172} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.744862] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df2a3fc3-f318-4102-8bf4-908ab36d5d29 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.751861] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 909.751861] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e75b35-afdd-7013-dca6-95c2d2248620" [ 909.751861] env[61867]: _type = "Task" [ 909.751861] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.760794] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e75b35-afdd-7013-dca6-95c2d2248620, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.780714] env[61867]: DEBUG oslo_concurrency.lockutils [req-9fcb3257-c081-4d7e-a389-4e6a4549938b req-1fdd0ffe-8711-4a6d-9d3d-32f30672380e service nova] Releasing lock "refresh_cache-5ec9eb1f-1d24-41d9-b5c7-a478a427c563" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.781040] env[61867]: DEBUG nova.compute.manager [req-9fcb3257-c081-4d7e-a389-4e6a4549938b req-1fdd0ffe-8711-4a6d-9d3d-32f30672380e service nova] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Received event network-vif-deleted-eaffa9e8-2616-4b49-9668-c8cc7ef884c0 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.826009] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 909.826297] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a144ba97-c779-413b-ac03-bfdae6276952 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.834031] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 909.834031] env[61867]: value = "task-1276836" [ 909.834031] env[61867]: _type = "Task" [ 909.834031] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.842502] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276836, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.933249] env[61867]: DEBUG oslo_vmware.api [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276828, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.089975] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276834, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.156565] env[61867]: DEBUG oslo_vmware.api [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276831, 'name': PowerOnVM_Task, 'duration_secs': 1.669992} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.156859] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 910.157186] env[61867]: INFO nova.compute.manager [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Took 12.34 seconds to spawn the instance on the hypervisor. [ 910.157401] env[61867]: DEBUG nova.compute.manager [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 910.158240] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f413aec-0f4b-47dd-9843-c6c452de2fee {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.236557] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a631d67d-db26-44fa-83d4-146ccf937112 tempest-AttachVolumeTestJSON-101616288 tempest-AttachVolumeTestJSON-101616288-project-member] Lock "6dda389d-a8c5-4e0e-87a5-4065e24c034e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.939s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.271128] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e75b35-afdd-7013-dca6-95c2d2248620, 'name': SearchDatastore_Task, 'duration_secs': 0.035514} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.271128] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.271128] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 5ec9eb1f-1d24-41d9-b5c7-a478a427c563/5ec9eb1f-1d24-41d9-b5c7-a478a427c563.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 910.271128] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-abd97563-a804-4380-8e1e-707f38c06610 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.281322] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 910.281322] env[61867]: value = "task-1276837" [ 910.281322] env[61867]: _type = "Task" [ 910.281322] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.291064] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276837, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.344462] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276836, 'name': PowerOffVM_Task, 'duration_secs': 0.394374} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.344800] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 910.345915] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9a2b778-3de5-4e80-a72e-d3f79b2e9b2d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.365517] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d3b826-5702-49f0-bfaa-e1144dd9fe2d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.436975] env[61867]: DEBUG oslo_vmware.api [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276828, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.590084] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276834, 'name': ReconfigVM_Task, 'duration_secs': 0.677043} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.590601] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 816ca796-d8ed-4843-9b1f-f169f48ff047/816ca796-d8ed-4843-9b1f-f169f48ff047.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.591451] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5bb5af26-4b32-418c-9d79-c6b37b77e7ea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.600345] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 910.600345] env[61867]: value = "task-1276838" [ 910.600345] env[61867]: _type = "Task" [ 910.600345] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.612105] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276838, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.686186] env[61867]: INFO nova.compute.manager [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Took 25.30 seconds to build instance. [ 910.802908] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276837, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.878596] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Creating Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 910.878968] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-096f0a83-479e-4009-9fae-760b295624fe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.891539] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 910.891539] env[61867]: value = "task-1276839" [ 910.891539] env[61867]: _type = "Task" [ 910.891539] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.901616] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276839, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.936963] env[61867]: DEBUG oslo_vmware.api [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276828, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.978460] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Volume attach. Driver type: vmdk {{(pid=61867) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 910.978614] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274387', 'volume_id': '87de05dc-cb88-42e1-a436-004b68dd4e5c', 'name': 'volume-87de05dc-cb88-42e1-a436-004b68dd4e5c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6d2dab88-4165-4952-8019-2eaf3b863115', 'attached_at': '', 'detached_at': '', 'volume_id': '87de05dc-cb88-42e1-a436-004b68dd4e5c', 'serial': '87de05dc-cb88-42e1-a436-004b68dd4e5c'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 910.979648] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d857cb6e-ecb0-4148-8876-cf6e7b32ce64 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.998384] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2a8f97-7eb4-475c-9faa-6a4307ac3bb7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.029686] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] volume-87de05dc-cb88-42e1-a436-004b68dd4e5c/volume-87de05dc-cb88-42e1-a436-004b68dd4e5c.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.030076] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-222d3632-17ef-405d-964d-af1308da25f2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.054873] env[61867]: DEBUG oslo_vmware.api [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 911.054873] env[61867]: value = "task-1276840" [ 911.054873] env[61867]: _type = "Task" [ 911.054873] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.065320] env[61867]: DEBUG oslo_vmware.api [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276840, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.118129] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276838, 'name': Rename_Task, 'duration_secs': 0.420316} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.118473] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 911.118811] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee5b9f6a-fc40-4280-a1e5-fc2963bb1163 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.127428] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 911.127428] env[61867]: value = "task-1276841" [ 911.127428] env[61867]: _type = "Task" [ 911.127428] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.140691] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276841, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.188933] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3ea273a5-9270-4b81-a3c1-45e365dcc017 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.813s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.291656] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276837, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.683996} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.292059] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 5ec9eb1f-1d24-41d9-b5c7-a478a427c563/5ec9eb1f-1d24-41d9-b5c7-a478a427c563.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 911.292213] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.292496] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0f140c78-ff00-4f45-bdbd-d0afdb730c53 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.300207] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 911.300207] env[61867]: value = "task-1276843" [ 911.300207] env[61867]: _type = "Task" [ 911.300207] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.308922] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276843, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.411778] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276839, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.436724] env[61867]: DEBUG oslo_vmware.api [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276828, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.572420] env[61867]: DEBUG oslo_vmware.api [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276840, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.639234] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276841, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.766942] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Volume attach. Driver type: vmdk {{(pid=61867) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 911.767309] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274389', 'volume_id': '5c23a596-074b-4f75-a4bc-86c77ffc5080', 'name': 'volume-5c23a596-074b-4f75-a4bc-86c77ffc5080', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2aa08603-d87f-4734-bdfe-fdd610d54e1f', 'attached_at': '', 'detached_at': '', 'volume_id': '5c23a596-074b-4f75-a4bc-86c77ffc5080', 'serial': '5c23a596-074b-4f75-a4bc-86c77ffc5080'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 911.768985] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0bb600d-7072-49e1-83f2-7386b4e6fbdb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.794214] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac8fbfc-45f0-4544-968f-7c1e4ed7c4eb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.825616] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] volume-5c23a596-074b-4f75-a4bc-86c77ffc5080/volume-5c23a596-074b-4f75-a4bc-86c77ffc5080.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.826391] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-521323ee-ce5a-49ce-8fa8-a971ddb3df36 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.844820] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276843, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072241} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.844820] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.845591] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743a80c0-fd3f-4fe6-9993-b467c2aeae84 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.850933] env[61867]: DEBUG oslo_vmware.api [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 911.850933] env[61867]: value = "task-1276844" [ 911.850933] env[61867]: _type = "Task" [ 911.850933] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.880372] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 5ec9eb1f-1d24-41d9-b5c7-a478a427c563/5ec9eb1f-1d24-41d9-b5c7-a478a427c563.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.881528] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12e20194-81cb-4fc2-93a6-ed688c3e856f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.905506] env[61867]: DEBUG oslo_vmware.api [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276844, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.915866] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276839, 'name': CreateSnapshot_Task, 'duration_secs': 0.888339} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.917836] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Created Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 911.918377] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 911.918377] env[61867]: value = "task-1276845" [ 911.918377] env[61867]: _type = "Task" [ 911.918377] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.919630] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ffe88c-8b7a-458e-ba31-7d3540c38622 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.941063] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276845, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.948396] env[61867]: DEBUG oslo_vmware.api [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276828, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.068482] env[61867]: DEBUG oslo_vmware.api [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276840, 'name': ReconfigVM_Task, 'duration_secs': 0.822753} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.069595] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Reconfigured VM instance instance-0000004c to attach disk [datastore2] volume-87de05dc-cb88-42e1-a436-004b68dd4e5c/volume-87de05dc-cb88-42e1-a436-004b68dd4e5c.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.080205] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6c486e4-b38e-49ea-99d6-231447696bc7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.101217] env[61867]: DEBUG oslo_vmware.api [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 912.101217] env[61867]: value = "task-1276846" [ 912.101217] env[61867]: _type = "Task" [ 912.101217] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.112586] env[61867]: DEBUG oslo_vmware.api [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276846, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.140685] env[61867]: DEBUG oslo_vmware.api [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276841, 'name': PowerOnVM_Task, 'duration_secs': 0.883115} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.141342] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 912.141819] env[61867]: INFO nova.compute.manager [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Took 8.98 seconds to spawn the instance on the hypervisor. [ 912.142148] env[61867]: DEBUG nova.compute.manager [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 912.143196] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d38feb3-27f1-4c32-9937-5f3fc210ec71 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.365515] env[61867]: DEBUG oslo_vmware.api [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276844, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.461491] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Creating linked-clone VM from snapshot {{(pid=61867) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 912.462338] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276845, 'name': ReconfigVM_Task, 'duration_secs': 0.310248} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.462338] env[61867]: DEBUG oslo_vmware.api [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276828, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.462338] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d25d1765-0b88-4bd2-83ae-f5ee630ef4e6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.465489] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 5ec9eb1f-1d24-41d9-b5c7-a478a427c563/5ec9eb1f-1d24-41d9-b5c7-a478a427c563.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.466169] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b34efd5c-4cfa-4f4b-8d84-4dc2ebc93cb9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.474956] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 912.474956] env[61867]: value = "task-1276848" [ 912.474956] env[61867]: _type = "Task" [ 912.474956] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.477427] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 912.477427] env[61867]: value = "task-1276847" [ 912.477427] env[61867]: _type = "Task" [ 912.477427] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.489370] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276848, 'name': CloneVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.492791] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276847, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.615953] env[61867]: DEBUG oslo_vmware.api [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276846, 'name': ReconfigVM_Task, 'duration_secs': 0.2361} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.616441] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274387', 'volume_id': '87de05dc-cb88-42e1-a436-004b68dd4e5c', 'name': 'volume-87de05dc-cb88-42e1-a436-004b68dd4e5c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6d2dab88-4165-4952-8019-2eaf3b863115', 'attached_at': '', 'detached_at': '', 'volume_id': '87de05dc-cb88-42e1-a436-004b68dd4e5c', 'serial': '87de05dc-cb88-42e1-a436-004b68dd4e5c'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 912.663380] env[61867]: INFO nova.compute.manager [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Took 13.89 seconds to build instance. [ 912.863722] env[61867]: DEBUG oslo_vmware.api [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276844, 'name': ReconfigVM_Task, 'duration_secs': 0.612137} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.864065] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Reconfigured VM instance instance-0000004d to attach disk [datastore2] volume-5c23a596-074b-4f75-a4bc-86c77ffc5080/volume-5c23a596-074b-4f75-a4bc-86c77ffc5080.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.869351] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db22706e-e474-47df-8d87-7cce30cda35e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.886790] env[61867]: DEBUG oslo_vmware.api [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 912.886790] env[61867]: value = "task-1276849" [ 912.886790] env[61867]: _type = "Task" [ 912.886790] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.897381] env[61867]: DEBUG oslo_vmware.api [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276849, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.943130] env[61867]: DEBUG oslo_vmware.api [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276828, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.997439] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276847, 'name': Rename_Task, 'duration_secs': 0.169934} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.002563] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 913.003060] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276848, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.003163] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2eeb2968-b417-4f92-8ea7-1fde4ee7d7db {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.012411] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 913.012411] env[61867]: value = "task-1276850" [ 913.012411] env[61867]: _type = "Task" [ 913.012411] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.022262] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276850, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.169930] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9834ccc0-e5bc-409b-89b0-67e7ba91dfa7 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "816ca796-d8ed-4843-9b1f-f169f48ff047" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.416s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.397585] env[61867]: DEBUG oslo_vmware.api [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276849, 'name': ReconfigVM_Task, 'duration_secs': 0.198156} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.398082] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274389', 'volume_id': '5c23a596-074b-4f75-a4bc-86c77ffc5080', 'name': 'volume-5c23a596-074b-4f75-a4bc-86c77ffc5080', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2aa08603-d87f-4734-bdfe-fdd610d54e1f', 'attached_at': '', 'detached_at': '', 'volume_id': '5c23a596-074b-4f75-a4bc-86c77ffc5080', 'serial': '5c23a596-074b-4f75-a4bc-86c77ffc5080'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 913.443449] env[61867]: DEBUG oslo_vmware.api [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276828, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.491535] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276848, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.524864] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276850, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.672355] env[61867]: DEBUG nova.objects.instance [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lazy-loading 'flavor' on Instance uuid 6d2dab88-4165-4952-8019-2eaf3b863115 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.783338] env[61867]: DEBUG nova.compute.manager [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Stashing vm_state: active {{(pid=61867) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 913.945957] env[61867]: DEBUG oslo_vmware.api [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276828, 'name': ReconfigVM_Task, 'duration_secs': 5.87847} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.948316] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.948659] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Reconfigured VM to detach interface {{(pid=61867) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 913.993248] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276848, 'name': CloneVM_Task} progress is 95%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.031364] env[61867]: DEBUG oslo_vmware.api [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276850, 'name': PowerOnVM_Task, 'duration_secs': 0.671294} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.031749] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 914.031890] env[61867]: INFO nova.compute.manager [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Took 8.52 seconds to spawn the instance on the hypervisor. [ 914.032085] env[61867]: DEBUG nova.compute.manager [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.033250] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335808a4-80da-442e-9961-2a302aafa3ae {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.178097] env[61867]: DEBUG oslo_concurrency.lockutils [None req-35331753-4487-4bd2-b046-21b2f80a864d tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "6d2dab88-4165-4952-8019-2eaf3b863115" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.868s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.309170] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.309170] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.452904] env[61867]: DEBUG nova.objects.instance [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lazy-loading 'flavor' on Instance uuid 2aa08603-d87f-4734-bdfe-fdd610d54e1f {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.493869] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276848, 'name': CloneVM_Task, 'duration_secs': 1.681126} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.494414] env[61867]: INFO nova.virt.vmwareapi.vmops [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Created linked-clone VM from snapshot [ 914.495667] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4bd9be6-a352-4242-9442-2a93a907de2e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.505272] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Uploading image b2039ceb-4b65-4991-a2c7-32276f4cea84 {{(pid=61867) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 914.535646] env[61867]: DEBUG oslo_vmware.rw_handles [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 914.535646] env[61867]: value = "vm-274392" [ 914.535646] env[61867]: _type = "VirtualMachine" [ 914.535646] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 914.535945] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e18b223d-a498-4241-87bc-ee3303192f37 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.546470] env[61867]: DEBUG oslo_vmware.rw_handles [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lease: (returnval){ [ 914.546470] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5228752d-fbaf-414d-caa2-040cb61b8083" [ 914.546470] env[61867]: _type = "HttpNfcLease" [ 914.546470] env[61867]: } obtained for exporting VM: (result){ [ 914.546470] env[61867]: value = "vm-274392" [ 914.546470] env[61867]: _type = "VirtualMachine" [ 914.546470] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 914.546821] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the lease: (returnval){ [ 914.546821] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5228752d-fbaf-414d-caa2-040cb61b8083" [ 914.546821] env[61867]: _type = "HttpNfcLease" [ 914.546821] env[61867]: } to be ready. {{(pid=61867) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 914.561826] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 914.561826] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5228752d-fbaf-414d-caa2-040cb61b8083" [ 914.561826] env[61867]: _type = "HttpNfcLease" [ 914.561826] env[61867]: } is ready. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 914.564799] env[61867]: INFO nova.compute.manager [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Took 15.11 seconds to build instance. [ 914.574321] env[61867]: DEBUG oslo_vmware.rw_handles [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 914.574321] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5228752d-fbaf-414d-caa2-040cb61b8083" [ 914.574321] env[61867]: _type = "HttpNfcLease" [ 914.574321] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 914.575967] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d524113c-61f0-4da4-b412-0a2a90aeb55b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.593668] env[61867]: DEBUG oslo_vmware.rw_handles [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5228c0e8-d98d-a403-54e5-35d927db97d1/disk-0.vmdk from lease info. {{(pid=61867) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 914.593668] env[61867]: DEBUG oslo_vmware.rw_handles [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5228c0e8-d98d-a403-54e5-35d927db97d1/disk-0.vmdk for reading. {{(pid=61867) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 914.699317] env[61867]: DEBUG nova.compute.manager [req-46ba2aca-c549-4b80-9563-d0a86f868dca req-326385f7-f134-48a5-9078-fca64d084b7f service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Received event network-vif-deleted-65189f98-8c73-47e1-94cd-60c94c3a65d7 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.699579] env[61867]: INFO nova.compute.manager [req-46ba2aca-c549-4b80-9563-d0a86f868dca req-326385f7-f134-48a5-9078-fca64d084b7f service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Neutron deleted interface 65189f98-8c73-47e1-94cd-60c94c3a65d7; detaching it from the instance and deleting it from the info cache [ 914.701456] env[61867]: DEBUG nova.network.neutron [req-46ba2aca-c549-4b80-9563-d0a86f868dca req-326385f7-f134-48a5-9078-fca64d084b7f service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Updating instance_info_cache with network_info: [{"id": "af89c243-c119-47c9-9eda-60c9be2b8c51", "address": "fa:16:3e:4d:1f:76", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf89c243-c1", "ovs_interfaceid": "af89c243-c119-47c9-9eda-60c9be2b8c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a3bb848c-f036-4379-9e73-e7089972da31", "address": "fa:16:3e:6a:34:45", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3bb848c-f0", "ovs_interfaceid": "a3bb848c-f036-4379-9e73-e7089972da31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.704857] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9d636a51-787c-48a8-8b26-e5e64ee903da {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.816269] env[61867]: INFO nova.compute.claims [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 914.964351] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b0663ca6-7cbc-4b3a-9d3f-246403051a81 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.844s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.075501] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c61192de-6dab-45db-b903-96600950441d tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "5ec9eb1f-1d24-41d9-b5c7-a478a427c563" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.636s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.206142] env[61867]: DEBUG oslo_concurrency.lockutils [req-46ba2aca-c549-4b80-9563-d0a86f868dca req-326385f7-f134-48a5-9078-fca64d084b7f service nova] Acquiring lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.206362] env[61867]: DEBUG oslo_concurrency.lockutils [req-46ba2aca-c549-4b80-9563-d0a86f868dca req-326385f7-f134-48a5-9078-fca64d084b7f service nova] Acquired lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.207324] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81fd66a7-9bc6-4e7a-a798-d9da1ba52e40 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.228605] env[61867]: DEBUG oslo_concurrency.lockutils [req-46ba2aca-c549-4b80-9563-d0a86f868dca req-326385f7-f134-48a5-9078-fca64d084b7f service nova] Releasing lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.229459] env[61867]: WARNING nova.compute.manager [req-46ba2aca-c549-4b80-9563-d0a86f868dca req-326385f7-f134-48a5-9078-fca64d084b7f service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Detach interface failed, port_id=65189f98-8c73-47e1-94cd-60c94c3a65d7, reason: No device with interface-id 65189f98-8c73-47e1-94cd-60c94c3a65d7 exists on VM: nova.exception.NotFound: No device with interface-id 65189f98-8c73-47e1-94cd-60c94c3a65d7 exists on VM [ 915.268340] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.268699] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.268959] env[61867]: DEBUG nova.network.neutron [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 915.305442] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "6d20f463-9198-4590-8ec2-db471c64ba7c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.305822] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "6d20f463-9198-4590-8ec2-db471c64ba7c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.322349] env[61867]: INFO nova.compute.resource_tracker [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Updating resource usage from migration 328e4a90-96f6-4e27-ba58-da709d0e6338 [ 915.407680] env[61867]: DEBUG oslo_concurrency.lockutils [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquiring lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.409017] env[61867]: DEBUG oslo_concurrency.lockutils [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.615594] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d8122f-f70b-45cd-a2c1-f7ebcfe7f1b7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.627951] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef5e727-af06-47bf-8c04-f524b92fc9ea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.663982] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d0f51a5-8f10-4431-aace-aef5c2d3ecad {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.678061] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c5c598-15cd-4a54-984e-02b6728470d0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.694788] env[61867]: DEBUG nova.compute.provider_tree [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 915.755609] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "6d2dab88-4165-4952-8019-2eaf3b863115" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.755877] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "6d2dab88-4165-4952-8019-2eaf3b863115" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.808571] env[61867]: DEBUG nova.compute.manager [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 915.857191] env[61867]: DEBUG oslo_vmware.rw_handles [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b55588-fd0c-3a5b-7ac3-14d186a85ed3/disk-0.vmdk. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 915.858230] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de05f20-31b5-4301-b3d7-73342b258643 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.868772] env[61867]: DEBUG oslo_vmware.rw_handles [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b55588-fd0c-3a5b-7ac3-14d186a85ed3/disk-0.vmdk is in state: ready. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 915.869114] env[61867]: ERROR oslo_vmware.rw_handles [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b55588-fd0c-3a5b-7ac3-14d186a85ed3/disk-0.vmdk due to incomplete transfer. [ 915.871863] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6c38ed46-6bbc-4a42-a533-7b36af477fdf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.881976] env[61867]: DEBUG oslo_vmware.rw_handles [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b55588-fd0c-3a5b-7ac3-14d186a85ed3/disk-0.vmdk. {{(pid=61867) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 915.882999] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Uploaded image e21e0ddf-319b-457b-b08e-49317c63fa31 to the Glance image server {{(pid=61867) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 915.884564] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Destroying the VM {{(pid=61867) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 915.886123] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-470d51ea-b259-4fbc-b8ce-fc2800bca5b9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.893963] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 915.893963] env[61867]: value = "task-1276853" [ 915.893963] env[61867]: _type = "Task" [ 915.893963] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.903759] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276853, 'name': Destroy_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.911836] env[61867]: DEBUG nova.compute.manager [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 916.061261] env[61867]: INFO nova.network.neutron [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Port a3bb848c-f036-4379-9e73-e7089972da31 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 916.063092] env[61867]: DEBUG nova.network.neutron [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Updating instance_info_cache with network_info: [{"id": "af89c243-c119-47c9-9eda-60c9be2b8c51", "address": "fa:16:3e:4d:1f:76", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf89c243-c1", "ovs_interfaceid": "af89c243-c119-47c9-9eda-60c9be2b8c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.200770] env[61867]: DEBUG nova.scheduler.client.report [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.260978] env[61867]: INFO nova.compute.manager [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Detaching volume 87de05dc-cb88-42e1-a436-004b68dd4e5c [ 916.300243] env[61867]: INFO nova.virt.block_device [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Attempting to driver detach volume 87de05dc-cb88-42e1-a436-004b68dd4e5c from mountpoint /dev/sdb [ 916.300502] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Volume detach. Driver type: vmdk {{(pid=61867) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 916.300689] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274387', 'volume_id': '87de05dc-cb88-42e1-a436-004b68dd4e5c', 'name': 'volume-87de05dc-cb88-42e1-a436-004b68dd4e5c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6d2dab88-4165-4952-8019-2eaf3b863115', 'attached_at': '', 'detached_at': '', 'volume_id': '87de05dc-cb88-42e1-a436-004b68dd4e5c', 'serial': '87de05dc-cb88-42e1-a436-004b68dd4e5c'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 916.301700] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce28fbd-1458-4f5d-9ddb-211511e88bb0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.304632] env[61867]: DEBUG oslo_concurrency.lockutils [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.304870] env[61867]: DEBUG oslo_concurrency.lockutils [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.305092] env[61867]: DEBUG oslo_concurrency.lockutils [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "ca2f6d18-f773-4875-83de-2f2be912c2f8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.305291] env[61867]: DEBUG oslo_concurrency.lockutils [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "ca2f6d18-f773-4875-83de-2f2be912c2f8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.305466] env[61867]: DEBUG oslo_concurrency.lockutils [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "ca2f6d18-f773-4875-83de-2f2be912c2f8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.308204] env[61867]: INFO nova.compute.manager [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Terminating instance [ 916.310279] env[61867]: DEBUG nova.compute.manager [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 916.310482] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 916.311290] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f4bb1e-9e41-4041-a8bb-a9183d6fd409 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.338015] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1411e89a-3b19-48fe-b1b1-a7d1381cb73f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.346085] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 916.347216] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88bf0a1b-0fe0-4fce-be37-123154a03271 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.353643] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aba798e-b359-4b53-bef7-dbcc6cfceb84 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.360369] env[61867]: DEBUG oslo_vmware.api [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 916.360369] env[61867]: value = "task-1276854" [ 916.360369] env[61867]: _type = "Task" [ 916.360369] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.386973] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.388924] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3caf358f-754f-4cfd-8dae-b26586361429 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.395370] env[61867]: DEBUG oslo_vmware.api [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276854, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.413657] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] The volume has not been displaced from its original location: [datastore2] volume-87de05dc-cb88-42e1-a436-004b68dd4e5c/volume-87de05dc-cb88-42e1-a436-004b68dd4e5c.vmdk. No consolidation needed. {{(pid=61867) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 916.420436] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Reconfiguring VM instance instance-0000004c to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 916.424287] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e78a8454-eefb-477c-9e7e-25b683ed0352 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.445320] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276853, 'name': Destroy_Task, 'duration_secs': 0.532586} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.447010] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Destroyed the VM [ 916.447010] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Deleting Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 916.447485] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8b1b6d53-d3f0-4f85-86bc-d25c4cc5acc8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.454909] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "5ec9eb1f-1d24-41d9-b5c7-a478a427c563" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.454909] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "5ec9eb1f-1d24-41d9-b5c7-a478a427c563" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.454909] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "5ec9eb1f-1d24-41d9-b5c7-a478a427c563-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.454909] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "5ec9eb1f-1d24-41d9-b5c7-a478a427c563-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.454909] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "5ec9eb1f-1d24-41d9-b5c7-a478a427c563-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.456618] env[61867]: DEBUG oslo_vmware.api [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 916.456618] env[61867]: value = "task-1276855" [ 916.456618] env[61867]: _type = "Task" [ 916.456618] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.457171] env[61867]: INFO nova.compute.manager [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Terminating instance [ 916.460336] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 916.460336] env[61867]: value = "task-1276856" [ 916.460336] env[61867]: _type = "Task" [ 916.460336] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.461076] env[61867]: DEBUG nova.compute.manager [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 916.461392] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 916.466316] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9fc676-e212-4eb4-875e-bfaea4892d5a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.470125] env[61867]: DEBUG oslo_concurrency.lockutils [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.483643] env[61867]: DEBUG oslo_vmware.api [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276855, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.484012] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276856, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.486499] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 916.486810] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01730405-5a58-4c9a-bea0-b5df6badc7db {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.501581] env[61867]: DEBUG oslo_vmware.api [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 916.501581] env[61867]: value = "task-1276857" [ 916.501581] env[61867]: _type = "Task" [ 916.501581] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.513917] env[61867]: DEBUG oslo_vmware.api [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276857, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.565499] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "refresh_cache-ca2f6d18-f773-4875-83de-2f2be912c2f8" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.707939] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.399s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.708298] env[61867]: INFO nova.compute.manager [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Migrating [ 916.717463] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.330s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.719706] env[61867]: INFO nova.compute.claims [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 916.745802] env[61867]: DEBUG nova.compute.manager [req-fbdaaf2e-5792-4966-bc45-cafd3e5bc756 req-7951cc3e-fe18-438e-9c86-ae8d0bc216b0 service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Received event network-vif-deleted-a3bb848c-f036-4379-9e73-e7089972da31 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.873563] env[61867]: DEBUG oslo_vmware.api [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276854, 'name': PowerOffVM_Task, 'duration_secs': 0.285313} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.874025] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 916.874228] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 916.874674] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b52cf99-a028-45f2-96f8-007516c66c94 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.973084] env[61867]: DEBUG oslo_vmware.api [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276855, 'name': ReconfigVM_Task, 'duration_secs': 0.32873} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.973936] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Reconfigured VM instance instance-0000004c to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 916.981963] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b412e5b-f33a-445c-b566-09a587570d2a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.992199] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276856, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.993633] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 916.993846] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 916.994051] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Deleting the datastore file [datastore1] ca2f6d18-f773-4875-83de-2f2be912c2f8 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.994703] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ee757457-a4c9-45f7-80c6-a24c9ae777fa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.000430] env[61867]: DEBUG oslo_vmware.api [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 917.000430] env[61867]: value = "task-1276859" [ 917.000430] env[61867]: _type = "Task" [ 917.000430] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.002025] env[61867]: DEBUG oslo_vmware.api [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 917.002025] env[61867]: value = "task-1276860" [ 917.002025] env[61867]: _type = "Task" [ 917.002025] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.018694] env[61867]: DEBUG oslo_vmware.api [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276859, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.025046] env[61867]: DEBUG oslo_vmware.api [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276857, 'name': PowerOffVM_Task, 'duration_secs': 0.281456} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.025294] env[61867]: DEBUG oslo_vmware.api [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276860, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.025553] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 917.025726] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 917.025991] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-864b53a4-f87e-40c1-b87b-faa0fc3aa3e2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.071887] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9ebbe74b-29ac-4faa-82cf-a5e862bb3ae4 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "interface-ca2f6d18-f773-4875-83de-2f2be912c2f8-65189f98-8c73-47e1-94cd-60c94c3a65d7" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.743s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.104374] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 917.104374] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 917.104374] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleting the datastore file [datastore2] 5ec9eb1f-1d24-41d9-b5c7-a478a427c563 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 917.104374] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-968c88bf-a461-43f9-9f58-8e452ec39f32 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.112922] env[61867]: DEBUG oslo_vmware.api [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 917.112922] env[61867]: value = "task-1276862" [ 917.112922] env[61867]: _type = "Task" [ 917.112922] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.132024] env[61867]: DEBUG oslo_vmware.api [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276862, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.233757] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "refresh_cache-705a32b6-67f9-42cc-b4d0-f6d1783c68b5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.233980] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "refresh_cache-705a32b6-67f9-42cc-b4d0-f6d1783c68b5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.234181] env[61867]: DEBUG nova.network.neutron [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 917.381065] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "8a7f4314-0fd4-49f9-8eb6-12baa0977a53" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.381377] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "8a7f4314-0fd4-49f9-8eb6-12baa0977a53" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.477287] env[61867]: DEBUG oslo_vmware.api [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276856, 'name': RemoveSnapshot_Task, 'duration_secs': 0.710748} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.477572] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Deleted Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 917.477808] env[61867]: INFO nova.compute.manager [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Took 17.26 seconds to snapshot the instance on the hypervisor. [ 917.517540] env[61867]: DEBUG oslo_vmware.api [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276859, 'name': ReconfigVM_Task, 'duration_secs': 0.205914} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.520987] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274387', 'volume_id': '87de05dc-cb88-42e1-a436-004b68dd4e5c', 'name': 'volume-87de05dc-cb88-42e1-a436-004b68dd4e5c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6d2dab88-4165-4952-8019-2eaf3b863115', 'attached_at': '', 'detached_at': '', 'volume_id': '87de05dc-cb88-42e1-a436-004b68dd4e5c', 'serial': '87de05dc-cb88-42e1-a436-004b68dd4e5c'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 917.523452] env[61867]: DEBUG oslo_vmware.api [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276860, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.387575} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.524418] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.524883] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 917.525111] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 917.525544] env[61867]: INFO nova.compute.manager [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Took 1.21 seconds to destroy the instance on the hypervisor. [ 917.525544] env[61867]: DEBUG oslo.service.loopingcall [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.525657] env[61867]: DEBUG nova.compute.manager [-] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 917.525749] env[61867]: DEBUG nova.network.neutron [-] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 917.592153] env[61867]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port a3bb848c-f036-4379-9e73-e7089972da31 could not be found.", "detail": ""}} {{(pid=61867) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 917.592153] env[61867]: DEBUG nova.network.neutron [-] Unable to show port a3bb848c-f036-4379-9e73-e7089972da31 as it no longer exists. {{(pid=61867) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 917.627150] env[61867]: DEBUG oslo_vmware.api [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276862, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.324088} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.627150] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.627150] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 917.627150] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 917.627150] env[61867]: INFO nova.compute.manager [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Took 1.17 seconds to destroy the instance on the hypervisor. [ 917.627150] env[61867]: DEBUG oslo.service.loopingcall [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.627150] env[61867]: DEBUG nova.compute.manager [-] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 917.627150] env[61867]: DEBUG nova.network.neutron [-] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 917.887400] env[61867]: DEBUG nova.compute.manager [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 917.998684] env[61867]: DEBUG nova.network.neutron [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Updating instance_info_cache with network_info: [{"id": "254a9643-f941-493e-8c87-2932a6cc00ce", "address": "fa:16:3e:85:01:b4", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap254a9643-f9", "ovs_interfaceid": "254a9643-f941-493e-8c87-2932a6cc00ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.051022] env[61867]: DEBUG nova.compute.manager [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Found 3 images (rotation: 2) {{(pid=61867) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 918.051022] env[61867]: DEBUG nova.compute.manager [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Rotating out 1 backups {{(pid=61867) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 918.051022] env[61867]: DEBUG nova.compute.manager [None req-ff4fcede-5666-42bc-a956-a43ee7d513ec tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Deleting image 53630313-ba7e-45a9-88e0-b3fbaaa31ae7 {{(pid=61867) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 918.060127] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a54dd5b-c628-452d-94b8-f0d253e76ab0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.068355] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ade57c-3980-43ab-bbb6-89bcb8f907f9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.113033] env[61867]: DEBUG nova.objects.instance [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lazy-loading 'flavor' on Instance uuid 6d2dab88-4165-4952-8019-2eaf3b863115 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 918.113824] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b571445d-d75d-4bae-801e-a6ebac24526f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.123913] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5151ecf-c08d-45d4-9150-cbfd47ec8cb3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.143757] env[61867]: DEBUG nova.compute.provider_tree [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.334471] env[61867]: DEBUG oslo_concurrency.lockutils [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "7479bf91-5aef-4e75-a127-7e82ae15a003" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.334801] env[61867]: DEBUG oslo_concurrency.lockutils [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "7479bf91-5aef-4e75-a127-7e82ae15a003" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.412640] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.506393] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "refresh_cache-705a32b6-67f9-42cc-b4d0-f6d1783c68b5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.646057] env[61867]: DEBUG nova.scheduler.client.report [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 918.757325] env[61867]: DEBUG nova.network.neutron [-] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.838329] env[61867]: DEBUG nova.compute.utils [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.892739] env[61867]: DEBUG nova.network.neutron [-] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.046958] env[61867]: DEBUG nova.compute.manager [req-d48dbb37-2015-443d-9fea-78a3e94a685b req-0d7f0f6e-6e29-4583-9f72-31e68fbac25a service nova] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Received event network-vif-deleted-f91d6c97-5188-4d51-8b56-411f6ca824c2 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.047081] env[61867]: DEBUG nova.compute.manager [req-d48dbb37-2015-443d-9fea-78a3e94a685b req-0d7f0f6e-6e29-4583-9f72-31e68fbac25a service nova] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Received event network-vif-deleted-af89c243-c119-47c9-9eda-60c9be2b8c51 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.123644] env[61867]: DEBUG oslo_concurrency.lockutils [None req-af611fb1-3345-4e28-97c3-241c1e686100 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "6d2dab88-4165-4952-8019-2eaf3b863115" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.366s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.154024] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.435s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.154024] env[61867]: DEBUG nova.compute.manager [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 919.155345] env[61867]: DEBUG oslo_concurrency.lockutils [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.685s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.157978] env[61867]: INFO nova.compute.claims [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 919.260327] env[61867]: INFO nova.compute.manager [-] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Took 1.63 seconds to deallocate network for instance. [ 919.341224] env[61867]: DEBUG oslo_concurrency.lockutils [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "7479bf91-5aef-4e75-a127-7e82ae15a003" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.397736] env[61867]: INFO nova.compute.manager [-] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Took 1.87 seconds to deallocate network for instance. [ 919.667101] env[61867]: DEBUG nova.compute.utils [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 919.668548] env[61867]: DEBUG nova.compute.manager [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 919.668670] env[61867]: DEBUG nova.network.neutron [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 919.710755] env[61867]: DEBUG nova.policy [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ab955c7bc494978bb235ef48300b5ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1dca1c24993044e499ef3ded17361258', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 919.766647] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.904955] env[61867]: DEBUG oslo_concurrency.lockutils [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.983793] env[61867]: DEBUG nova.network.neutron [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Successfully created port: 690a5747-8fcf-445f-9c80-198bfc4fef9a {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 920.028022] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4120a84d-40a6-41fb-9352-dee0c52702cb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.048657] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Updating instance '705a32b6-67f9-42cc-b4d0-f6d1783c68b5' progress to 0 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 920.138795] env[61867]: DEBUG oslo_concurrency.lockutils [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "3905193e-04da-439a-bf6c-16f638a692bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.139051] env[61867]: DEBUG oslo_concurrency.lockutils [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "3905193e-04da-439a-bf6c-16f638a692bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.172289] env[61867]: DEBUG nova.compute.manager [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 920.422500] env[61867]: DEBUG oslo_concurrency.lockutils [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "7479bf91-5aef-4e75-a127-7e82ae15a003" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.422573] env[61867]: DEBUG oslo_concurrency.lockutils [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "7479bf91-5aef-4e75-a127-7e82ae15a003" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.422840] env[61867]: INFO nova.compute.manager [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Attaching volume 9bcdabad-faa3-4b43-ac62-c59cb70802e4 to /dev/sdb [ 920.426876] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb093fb3-901c-43ba-a836-0c26092efa1c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.438138] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-900cfded-196c-494e-92bb-0bf8b92398b1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.483184] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe12afb-7884-4f17-9a24-aeb8b5eed561 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.487831] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33291bf3-8eb0-4d6d-9dc9-30912ca1ec82 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.498607] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c20c55d-3cac-4720-93cd-b7b1e9c78ac6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.503090] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d56c201-2cd2-4a01-8003-ed7fa6209045 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.515995] env[61867]: DEBUG nova.compute.provider_tree [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.521233] env[61867]: DEBUG nova.virt.block_device [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Updating existing volume attachment record: d77c2300-1a08-48ed-9298-37b034cc4f17 {{(pid=61867) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 920.555171] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 920.555498] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f92c2b4e-444d-4e6c-a85f-c3f154a1560e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.564354] env[61867]: DEBUG oslo_vmware.api [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 920.564354] env[61867]: value = "task-1276863" [ 920.564354] env[61867]: _type = "Task" [ 920.564354] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.574189] env[61867]: DEBUG oslo_vmware.api [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276863, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.641793] env[61867]: DEBUG nova.compute.manager [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 921.026074] env[61867]: DEBUG nova.scheduler.client.report [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 921.078353] env[61867]: DEBUG oslo_vmware.api [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276863, 'name': PowerOffVM_Task, 'duration_secs': 0.248042} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.079053] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 921.079490] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Updating instance '705a32b6-67f9-42cc-b4d0-f6d1783c68b5' progress to 17 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 921.094292] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bc160710-ed6e-43fb-9108-0bc8c36a03be tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.094661] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bc160710-ed6e-43fb-9108-0bc8c36a03be tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.094931] env[61867]: DEBUG nova.compute.manager [None req-bc160710-ed6e-43fb-9108-0bc8c36a03be tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 921.095925] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e58f14-a893-4e67-989d-8f0eab79a023 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.104529] env[61867]: DEBUG nova.compute.manager [None req-bc160710-ed6e-43fb-9108-0bc8c36a03be tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61867) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 921.105282] env[61867]: DEBUG nova.objects.instance [None req-bc160710-ed6e-43fb-9108-0bc8c36a03be tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lazy-loading 'flavor' on Instance uuid abb41c0c-6d0d-4147-a4af-554ab7d9e921 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.160970] env[61867]: DEBUG oslo_concurrency.lockutils [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.183978] env[61867]: DEBUG nova.compute.manager [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 921.207127] env[61867]: DEBUG nova.virt.hardware [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 921.207399] env[61867]: DEBUG nova.virt.hardware [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 921.207576] env[61867]: DEBUG nova.virt.hardware [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 921.207746] env[61867]: DEBUG nova.virt.hardware [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 921.207897] env[61867]: DEBUG nova.virt.hardware [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 921.208064] env[61867]: DEBUG nova.virt.hardware [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 921.208280] env[61867]: DEBUG nova.virt.hardware [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 921.208445] env[61867]: DEBUG nova.virt.hardware [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 921.208611] env[61867]: DEBUG nova.virt.hardware [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 921.208777] env[61867]: DEBUG nova.virt.hardware [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 921.208949] env[61867]: DEBUG nova.virt.hardware [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 921.209867] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb963bc-6c90-4ea9-b777-ed3beab9b356 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.218684] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e7dd1a-9ec1-40d5-8325-ce68a19bf01e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.440727] env[61867]: DEBUG nova.compute.manager [req-2b3471a6-bf57-4c70-a075-9d0dfc230c81 req-d084fc3d-fa4f-4581-aa57-365549378da3 service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Received event network-vif-plugged-690a5747-8fcf-445f-9c80-198bfc4fef9a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.441060] env[61867]: DEBUG oslo_concurrency.lockutils [req-2b3471a6-bf57-4c70-a075-9d0dfc230c81 req-d084fc3d-fa4f-4581-aa57-365549378da3 service nova] Acquiring lock "6d20f463-9198-4590-8ec2-db471c64ba7c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.441377] env[61867]: DEBUG oslo_concurrency.lockutils [req-2b3471a6-bf57-4c70-a075-9d0dfc230c81 req-d084fc3d-fa4f-4581-aa57-365549378da3 service nova] Lock "6d20f463-9198-4590-8ec2-db471c64ba7c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.441953] env[61867]: DEBUG oslo_concurrency.lockutils [req-2b3471a6-bf57-4c70-a075-9d0dfc230c81 req-d084fc3d-fa4f-4581-aa57-365549378da3 service nova] Lock "6d20f463-9198-4590-8ec2-db471c64ba7c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.442251] env[61867]: DEBUG nova.compute.manager [req-2b3471a6-bf57-4c70-a075-9d0dfc230c81 req-d084fc3d-fa4f-4581-aa57-365549378da3 service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] No waiting events found dispatching network-vif-plugged-690a5747-8fcf-445f-9c80-198bfc4fef9a {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 921.442542] env[61867]: WARNING nova.compute.manager [req-2b3471a6-bf57-4c70-a075-9d0dfc230c81 req-d084fc3d-fa4f-4581-aa57-365549378da3 service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Received unexpected event network-vif-plugged-690a5747-8fcf-445f-9c80-198bfc4fef9a for instance with vm_state building and task_state spawning. [ 921.533216] env[61867]: DEBUG oslo_concurrency.lockutils [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.376s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.533216] env[61867]: DEBUG nova.compute.manager [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 921.538515] env[61867]: DEBUG nova.network.neutron [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Successfully updated port: 690a5747-8fcf-445f-9c80-198bfc4fef9a {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 921.539725] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.127s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.541834] env[61867]: INFO nova.compute.claims [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 921.587217] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 921.587511] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 921.587701] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 921.587937] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 921.588088] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 921.588287] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 921.588591] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 921.588778] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 921.589400] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 921.589400] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 921.589400] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 921.595179] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a46c9546-d44e-4c50-8479-811835fd2672 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.609862] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc160710-ed6e-43fb-9108-0bc8c36a03be tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 921.610150] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5294a500-d171-4283-9045-820490c3a0fc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.614986] env[61867]: DEBUG oslo_vmware.api [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 921.614986] env[61867]: value = "task-1276867" [ 921.614986] env[61867]: _type = "Task" [ 921.614986] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.620621] env[61867]: DEBUG oslo_vmware.api [None req-bc160710-ed6e-43fb-9108-0bc8c36a03be tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 921.620621] env[61867]: value = "task-1276868" [ 921.620621] env[61867]: _type = "Task" [ 921.620621] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.631204] env[61867]: DEBUG oslo_vmware.api [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276867, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.638947] env[61867]: DEBUG oslo_vmware.api [None req-bc160710-ed6e-43fb-9108-0bc8c36a03be tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276868, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.046483] env[61867]: DEBUG nova.compute.utils [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 922.049565] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "refresh_cache-6d20f463-9198-4590-8ec2-db471c64ba7c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.049972] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquired lock "refresh_cache-6d20f463-9198-4590-8ec2-db471c64ba7c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.050308] env[61867]: DEBUG nova.network.neutron [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 922.054277] env[61867]: DEBUG nova.compute.manager [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 922.054656] env[61867]: DEBUG nova.network.neutron [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 922.107128] env[61867]: DEBUG nova.policy [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5a066048582d4f7d905e201884470509', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '60b35760e3e14245aea3600d36c838dc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 922.142365] env[61867]: DEBUG oslo_vmware.api [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276867, 'name': ReconfigVM_Task, 'duration_secs': 0.291662} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.142989] env[61867]: DEBUG oslo_vmware.api [None req-bc160710-ed6e-43fb-9108-0bc8c36a03be tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276868, 'name': PowerOffVM_Task, 'duration_secs': 0.32146} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.143478] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Updating instance '705a32b6-67f9-42cc-b4d0-f6d1783c68b5' progress to 33 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 922.147550] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc160710-ed6e-43fb-9108-0bc8c36a03be tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 922.147950] env[61867]: DEBUG nova.compute.manager [None req-bc160710-ed6e-43fb-9108-0bc8c36a03be tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 922.149234] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d46ee2e-a6a5-43d6-b2ff-c316e86b76bb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.382855] env[61867]: DEBUG nova.network.neutron [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Successfully created port: d9630611-bb16-428c-ade6-78ff0355d81d {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 922.555452] env[61867]: DEBUG nova.compute.manager [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 922.607548] env[61867]: DEBUG nova.network.neutron [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 922.651446] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 922.651446] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 922.651741] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 922.651777] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 922.652026] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 922.652351] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 922.652680] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 922.652903] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 922.653131] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 922.653343] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 922.653681] env[61867]: DEBUG nova.virt.hardware [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 922.658901] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Reconfiguring VM instance instance-00000054 to detach disk 2000 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 922.661916] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c60a0b37-6782-4d20-9961-99b0256d3083 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.679733] env[61867]: DEBUG nova.network.neutron [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Successfully created port: 017138c2-47c5-4033-a0b5-42707d74a494 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 922.682032] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bc160710-ed6e-43fb-9108-0bc8c36a03be tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.587s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.688366] env[61867]: DEBUG oslo_vmware.api [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 922.688366] env[61867]: value = "task-1276869" [ 922.688366] env[61867]: _type = "Task" [ 922.688366] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.700747] env[61867]: DEBUG oslo_vmware.api [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276869, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.851055] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3cd0789-6250-4b3e-a290-0b80f79294a4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.860509] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff16691-2cfd-4767-8ffc-0faa5b4943ac {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.893903] env[61867]: DEBUG nova.network.neutron [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Updating instance_info_cache with network_info: [{"id": "690a5747-8fcf-445f-9c80-198bfc4fef9a", "address": "fa:16:3e:99:85:22", "network": {"id": "d8a081a0-0d77-4ae7-9d58-19da456f7699", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-201249363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dca1c24993044e499ef3ded17361258", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap690a5747-8f", "ovs_interfaceid": "690a5747-8fcf-445f-9c80-198bfc4fef9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.897146] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8344493d-22d2-431f-a751-380cd9d34eab {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.904718] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb44f7f-0294-4a5b-b1b4-5bf8fbf329fb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.921077] env[61867]: DEBUG nova.compute.provider_tree [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.052410] env[61867]: DEBUG nova.network.neutron [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Successfully created port: 74ad152e-530d-49fd-bdc1-30e366f5f9ac {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 923.203354] env[61867]: DEBUG oslo_vmware.api [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276869, 'name': ReconfigVM_Task, 'duration_secs': 0.272977} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.203797] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Reconfigured VM instance instance-00000054 to detach disk 2000 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 923.205167] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6d4a8a-0b30-4d65-b85f-78d2af0d074f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.235547] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 705a32b6-67f9-42cc-b4d0-f6d1783c68b5/705a32b6-67f9-42cc-b4d0-f6d1783c68b5.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.236452] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9691e1a4-770a-4dd6-b6e6-cac341aa2a1e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.263482] env[61867]: DEBUG oslo_vmware.api [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 923.263482] env[61867]: value = "task-1276871" [ 923.263482] env[61867]: _type = "Task" [ 923.263482] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.275436] env[61867]: DEBUG oslo_vmware.api [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276871, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.399753] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Releasing lock "refresh_cache-6d20f463-9198-4590-8ec2-db471c64ba7c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.400177] env[61867]: DEBUG nova.compute.manager [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Instance network_info: |[{"id": "690a5747-8fcf-445f-9c80-198bfc4fef9a", "address": "fa:16:3e:99:85:22", "network": {"id": "d8a081a0-0d77-4ae7-9d58-19da456f7699", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-201249363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dca1c24993044e499ef3ded17361258", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap690a5747-8f", "ovs_interfaceid": "690a5747-8fcf-445f-9c80-198bfc4fef9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 923.400694] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:85:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '690a5747-8fcf-445f-9c80-198bfc4fef9a', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 923.410602] env[61867]: DEBUG oslo.service.loopingcall [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.410929] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 923.411229] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98ee34cf-b3c8-4489-af15-0b165538f501 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.431229] env[61867]: DEBUG nova.scheduler.client.report [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 923.447269] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 923.447269] env[61867]: value = "task-1276872" [ 923.447269] env[61867]: _type = "Task" [ 923.447269] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.459786] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276872, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.473258] env[61867]: DEBUG nova.compute.manager [req-8bb2e354-7f03-4412-b9f3-68b5c7fa4fc1 req-4df2cc95-d2d4-4177-ae9a-c83b8dda168a service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Received event network-changed-690a5747-8fcf-445f-9c80-198bfc4fef9a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.473335] env[61867]: DEBUG nova.compute.manager [req-8bb2e354-7f03-4412-b9f3-68b5c7fa4fc1 req-4df2cc95-d2d4-4177-ae9a-c83b8dda168a service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Refreshing instance network info cache due to event network-changed-690a5747-8fcf-445f-9c80-198bfc4fef9a. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 923.473576] env[61867]: DEBUG oslo_concurrency.lockutils [req-8bb2e354-7f03-4412-b9f3-68b5c7fa4fc1 req-4df2cc95-d2d4-4177-ae9a-c83b8dda168a service nova] Acquiring lock "refresh_cache-6d20f463-9198-4590-8ec2-db471c64ba7c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.473729] env[61867]: DEBUG oslo_concurrency.lockutils [req-8bb2e354-7f03-4412-b9f3-68b5c7fa4fc1 req-4df2cc95-d2d4-4177-ae9a-c83b8dda168a service nova] Acquired lock "refresh_cache-6d20f463-9198-4590-8ec2-db471c64ba7c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.474663] env[61867]: DEBUG nova.network.neutron [req-8bb2e354-7f03-4412-b9f3-68b5c7fa4fc1 req-4df2cc95-d2d4-4177-ae9a-c83b8dda168a service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Refreshing network info cache for port 690a5747-8fcf-445f-9c80-198bfc4fef9a {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 923.570437] env[61867]: DEBUG nova.compute.manager [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 923.605742] env[61867]: DEBUG nova.virt.hardware [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 923.606151] env[61867]: DEBUG nova.virt.hardware [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 923.606430] env[61867]: DEBUG nova.virt.hardware [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 923.606721] env[61867]: DEBUG nova.virt.hardware [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 923.606916] env[61867]: DEBUG nova.virt.hardware [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 923.607119] env[61867]: DEBUG nova.virt.hardware [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 923.607393] env[61867]: DEBUG nova.virt.hardware [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 923.607601] env[61867]: DEBUG nova.virt.hardware [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 923.607817] env[61867]: DEBUG nova.virt.hardware [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 923.608043] env[61867]: DEBUG nova.virt.hardware [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 923.608276] env[61867]: DEBUG nova.virt.hardware [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 923.610370] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead91e8d-2fbb-49b5-8db1-8e0d2d97d310 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.622521] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3260e51-29bc-411f-8d1d-0b4eb8c2902c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.765227] env[61867]: DEBUG oslo_vmware.rw_handles [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5228c0e8-d98d-a403-54e5-35d927db97d1/disk-0.vmdk. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 923.766354] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a14356e-d3f4-4edb-bad7-3865c52e0444 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.776384] env[61867]: DEBUG oslo_vmware.rw_handles [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5228c0e8-d98d-a403-54e5-35d927db97d1/disk-0.vmdk is in state: ready. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 923.776550] env[61867]: ERROR oslo_vmware.rw_handles [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5228c0e8-d98d-a403-54e5-35d927db97d1/disk-0.vmdk due to incomplete transfer. [ 923.779512] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c759d0f6-10f0-4437-aa9c-943a291e04a5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.781570] env[61867]: DEBUG oslo_vmware.api [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276871, 'name': ReconfigVM_Task, 'duration_secs': 0.297204} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.781889] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 705a32b6-67f9-42cc-b4d0-f6d1783c68b5/705a32b6-67f9-42cc-b4d0-f6d1783c68b5.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.782212] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Updating instance '705a32b6-67f9-42cc-b4d0-f6d1783c68b5' progress to 50 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 923.791369] env[61867]: DEBUG oslo_vmware.rw_handles [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5228c0e8-d98d-a403-54e5-35d927db97d1/disk-0.vmdk. {{(pid=61867) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 923.791369] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Uploaded image b2039ceb-4b65-4991-a2c7-32276f4cea84 to the Glance image server {{(pid=61867) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 923.793635] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Destroying the VM {{(pid=61867) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 923.793912] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ab52a058-1b35-45b3-a040-ddb5e23e5d5c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.801455] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 923.801455] env[61867]: value = "task-1276873" [ 923.801455] env[61867]: _type = "Task" [ 923.801455] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.810309] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276873, 'name': Destroy_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.938027] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.398s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.938418] env[61867]: DEBUG nova.compute.manager [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 923.945310] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.179s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.945547] env[61867]: DEBUG nova.objects.instance [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lazy-loading 'resources' on Instance uuid 5ec9eb1f-1d24-41d9-b5c7-a478a427c563 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.958017] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276872, 'name': CreateVM_Task, 'duration_secs': 0.372135} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.958689] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 923.959441] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.959617] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.959949] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 923.960425] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab722917-c1cf-49d4-9419-29b3c3128b17 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.965456] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 923.965456] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d3d986-e4c0-443a-a18c-b6e5f95fbdd2" [ 923.965456] env[61867]: _type = "Task" [ 923.965456] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.974957] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d3d986-e4c0-443a-a18c-b6e5f95fbdd2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.198379] env[61867]: DEBUG nova.network.neutron [req-8bb2e354-7f03-4412-b9f3-68b5c7fa4fc1 req-4df2cc95-d2d4-4177-ae9a-c83b8dda168a service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Updated VIF entry in instance network info cache for port 690a5747-8fcf-445f-9c80-198bfc4fef9a. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 924.198829] env[61867]: DEBUG nova.network.neutron [req-8bb2e354-7f03-4412-b9f3-68b5c7fa4fc1 req-4df2cc95-d2d4-4177-ae9a-c83b8dda168a service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Updating instance_info_cache with network_info: [{"id": "690a5747-8fcf-445f-9c80-198bfc4fef9a", "address": "fa:16:3e:99:85:22", "network": {"id": "d8a081a0-0d77-4ae7-9d58-19da456f7699", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-201249363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dca1c24993044e499ef3ded17361258", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap690a5747-8f", "ovs_interfaceid": "690a5747-8fcf-445f-9c80-198bfc4fef9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.288971] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91353d48-4e4a-43e5-8f99-b8d3039f9864 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.311821] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e5fd41-db86-4a29-8374-8ca5433228ec {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.321009] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276873, 'name': Destroy_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.334021] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Updating instance '705a32b6-67f9-42cc-b4d0-f6d1783c68b5' progress to 67 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 924.447139] env[61867]: DEBUG nova.compute.utils [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 924.450551] env[61867]: DEBUG nova.compute.manager [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 924.450727] env[61867]: DEBUG nova.network.neutron [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 924.457549] env[61867]: DEBUG nova.compute.manager [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Stashing vm_state: stopped {{(pid=61867) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 924.479091] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d3d986-e4c0-443a-a18c-b6e5f95fbdd2, 'name': SearchDatastore_Task, 'duration_secs': 0.009887} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.479719] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.480959] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 924.480959] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.480959] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.480959] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 924.481199] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f0551e00-61bb-459a-bdc8-d6add2dffc4a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.494235] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 924.494463] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 924.495217] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57e7f378-f0b1-4b80-97ff-dee7306ea494 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.501789] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 924.501789] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e30024-6067-a03d-e02b-98ab51e2b331" [ 924.501789] env[61867]: _type = "Task" [ 924.501789] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.515469] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e30024-6067-a03d-e02b-98ab51e2b331, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.522579] env[61867]: DEBUG nova.policy [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '322fe9fc12a34183b4241e28f7254df4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24180ce8aca142fb897e29d4853c20c3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 924.692478] env[61867]: DEBUG nova.network.neutron [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Successfully updated port: d9630611-bb16-428c-ade6-78ff0355d81d {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 924.702502] env[61867]: DEBUG oslo_concurrency.lockutils [req-8bb2e354-7f03-4412-b9f3-68b5c7fa4fc1 req-4df2cc95-d2d4-4177-ae9a-c83b8dda168a service nova] Releasing lock "refresh_cache-6d20f463-9198-4590-8ec2-db471c64ba7c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.743540] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54aa4fef-09af-4149-a122-b1c4fa1dba50 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.751712] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6cb09c-7656-49d7-9d86-37302167b15b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.785183] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6cb825-9adc-4aad-8823-010489d204b6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.791768] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e456d340-773d-49fd-ae26-ca5ee8d270ce {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.806710] env[61867]: DEBUG nova.compute.provider_tree [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.817739] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276873, 'name': Destroy_Task, 'duration_secs': 0.744418} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.817739] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Destroyed the VM [ 924.817968] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Deleting Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 924.818231] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e24f4b04-bd00-4bea-a5c0-bf0f022d66b3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.826903] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 924.826903] env[61867]: value = "task-1276874" [ 924.826903] env[61867]: _type = "Task" [ 924.826903] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.836635] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276874, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.896843] env[61867]: DEBUG nova.network.neutron [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Port 254a9643-f941-493e-8c87-2932a6cc00ce binding to destination host cpu-1 is already ACTIVE {{(pid=61867) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 924.918935] env[61867]: DEBUG nova.network.neutron [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Successfully created port: 211a1ec3-d618-4c1d-ac61-7e7b5681b9ed {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 924.954226] env[61867]: DEBUG nova.compute.manager [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 924.988229] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.017655] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e30024-6067-a03d-e02b-98ab51e2b331, 'name': SearchDatastore_Task, 'duration_secs': 0.014917} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.018523] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-357fe439-abc7-407e-89a1-10736af2d3b7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.024705] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 925.024705] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525f4315-6ffa-33e3-ab8b-c8aa91db0c4b" [ 925.024705] env[61867]: _type = "Task" [ 925.024705] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.033342] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525f4315-6ffa-33e3-ab8b-c8aa91db0c4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.074575] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Volume attach. Driver type: vmdk {{(pid=61867) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 925.074575] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274394', 'volume_id': '9bcdabad-faa3-4b43-ac62-c59cb70802e4', 'name': 'volume-9bcdabad-faa3-4b43-ac62-c59cb70802e4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7479bf91-5aef-4e75-a127-7e82ae15a003', 'attached_at': '', 'detached_at': '', 'volume_id': '9bcdabad-faa3-4b43-ac62-c59cb70802e4', 'serial': '9bcdabad-faa3-4b43-ac62-c59cb70802e4'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 925.075152] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781979b9-7909-444d-953c-4d284ae04e9b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.094166] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de667c0c-e2bb-4205-a6bb-9a3e2be14328 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.120135] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] volume-9bcdabad-faa3-4b43-ac62-c59cb70802e4/volume-9bcdabad-faa3-4b43-ac62-c59cb70802e4.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.120470] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d12da664-46b4-4c27-8211-a1ecb50225b2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.141591] env[61867]: DEBUG oslo_vmware.api [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 925.141591] env[61867]: value = "task-1276875" [ 925.141591] env[61867]: _type = "Task" [ 925.141591] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.150647] env[61867]: DEBUG oslo_vmware.api [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276875, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.309428] env[61867]: DEBUG nova.scheduler.client.report [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.339690] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276874, 'name': RemoveSnapshot_Task, 'duration_secs': 0.313486} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.339999] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Deleted Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 925.340500] env[61867]: DEBUG nova.compute.manager [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.341271] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34cffa51-240a-4cd6-b900-b2c71e2d752b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.506676] env[61867]: DEBUG nova.compute.manager [req-d512a7a3-cf9f-46be-bdbc-d37576859509 req-2d974286-e29c-4d60-bcb5-d3676c6e6037 service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Received event network-vif-plugged-d9630611-bb16-428c-ade6-78ff0355d81d {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.506921] env[61867]: DEBUG oslo_concurrency.lockutils [req-d512a7a3-cf9f-46be-bdbc-d37576859509 req-2d974286-e29c-4d60-bcb5-d3676c6e6037 service nova] Acquiring lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.507178] env[61867]: DEBUG oslo_concurrency.lockutils [req-d512a7a3-cf9f-46be-bdbc-d37576859509 req-2d974286-e29c-4d60-bcb5-d3676c6e6037 service nova] Lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.507326] env[61867]: DEBUG oslo_concurrency.lockutils [req-d512a7a3-cf9f-46be-bdbc-d37576859509 req-2d974286-e29c-4d60-bcb5-d3676c6e6037 service nova] Lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.507499] env[61867]: DEBUG nova.compute.manager [req-d512a7a3-cf9f-46be-bdbc-d37576859509 req-2d974286-e29c-4d60-bcb5-d3676c6e6037 service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] No waiting events found dispatching network-vif-plugged-d9630611-bb16-428c-ade6-78ff0355d81d {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 925.507666] env[61867]: WARNING nova.compute.manager [req-d512a7a3-cf9f-46be-bdbc-d37576859509 req-2d974286-e29c-4d60-bcb5-d3676c6e6037 service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Received unexpected event network-vif-plugged-d9630611-bb16-428c-ade6-78ff0355d81d for instance with vm_state building and task_state spawning. [ 925.507827] env[61867]: DEBUG nova.compute.manager [req-d512a7a3-cf9f-46be-bdbc-d37576859509 req-2d974286-e29c-4d60-bcb5-d3676c6e6037 service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Received event network-changed-d9630611-bb16-428c-ade6-78ff0355d81d {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.507986] env[61867]: DEBUG nova.compute.manager [req-d512a7a3-cf9f-46be-bdbc-d37576859509 req-2d974286-e29c-4d60-bcb5-d3676c6e6037 service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Refreshing instance network info cache due to event network-changed-d9630611-bb16-428c-ade6-78ff0355d81d. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 925.508190] env[61867]: DEBUG oslo_concurrency.lockutils [req-d512a7a3-cf9f-46be-bdbc-d37576859509 req-2d974286-e29c-4d60-bcb5-d3676c6e6037 service nova] Acquiring lock "refresh_cache-38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.508331] env[61867]: DEBUG oslo_concurrency.lockutils [req-d512a7a3-cf9f-46be-bdbc-d37576859509 req-2d974286-e29c-4d60-bcb5-d3676c6e6037 service nova] Acquired lock "refresh_cache-38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.508489] env[61867]: DEBUG nova.network.neutron [req-d512a7a3-cf9f-46be-bdbc-d37576859509 req-2d974286-e29c-4d60-bcb5-d3676c6e6037 service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Refreshing network info cache for port d9630611-bb16-428c-ade6-78ff0355d81d {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 925.537381] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525f4315-6ffa-33e3-ab8b-c8aa91db0c4b, 'name': SearchDatastore_Task, 'duration_secs': 0.0089} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.537381] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.537587] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 6d20f463-9198-4590-8ec2-db471c64ba7c/6d20f463-9198-4590-8ec2-db471c64ba7c.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 925.537852] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b50c9f55-58fa-44f4-9971-0b2c43937913 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.546297] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 925.546297] env[61867]: value = "task-1276876" [ 925.546297] env[61867]: _type = "Task" [ 925.546297] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.556149] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276876, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.651634] env[61867]: DEBUG oslo_vmware.api [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276875, 'name': ReconfigVM_Task, 'duration_secs': 0.391423} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.658021] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Reconfigured VM instance instance-00000050 to attach disk [datastore2] volume-9bcdabad-faa3-4b43-ac62-c59cb70802e4/volume-9bcdabad-faa3-4b43-ac62-c59cb70802e4.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 925.659967] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f81b4c76-69ed-4d73-b809-aff002393ad3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.682775] env[61867]: DEBUG oslo_vmware.api [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 925.682775] env[61867]: value = "task-1276877" [ 925.682775] env[61867]: _type = "Task" [ 925.682775] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.691955] env[61867]: DEBUG oslo_vmware.api [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276877, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.814221] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.869s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.817062] env[61867]: DEBUG oslo_concurrency.lockutils [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.912s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.817062] env[61867]: DEBUG nova.objects.instance [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lazy-loading 'resources' on Instance uuid ca2f6d18-f773-4875-83de-2f2be912c2f8 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.842118] env[61867]: INFO nova.scheduler.client.report [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleted allocations for instance 5ec9eb1f-1d24-41d9-b5c7-a478a427c563 [ 925.855008] env[61867]: INFO nova.compute.manager [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Shelve offloading [ 925.857035] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 925.858488] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-250e95c4-d102-46ef-9168-551b32fe21ea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.870864] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 925.870864] env[61867]: value = "task-1276878" [ 925.870864] env[61867]: _type = "Task" [ 925.870864] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.882980] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] VM already powered off {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 925.883311] env[61867]: DEBUG nova.compute.manager [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.884166] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9b3347-7c50-4c94-9050-d2d4200c61d9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.893538] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "refresh_cache-9e1cbfa9-28e9-4bca-adfd-78bf25428106" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.893980] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "refresh_cache-9e1cbfa9-28e9-4bca-adfd-78bf25428106" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.894260] env[61867]: DEBUG nova.network.neutron [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 925.917225] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.917481] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.917665] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.967212] env[61867]: DEBUG nova.compute.manager [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 925.992632] env[61867]: DEBUG nova.virt.hardware [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 925.992909] env[61867]: DEBUG nova.virt.hardware [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 925.993085] env[61867]: DEBUG nova.virt.hardware [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 925.993279] env[61867]: DEBUG nova.virt.hardware [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 925.993429] env[61867]: DEBUG nova.virt.hardware [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 925.993578] env[61867]: DEBUG nova.virt.hardware [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 925.993792] env[61867]: DEBUG nova.virt.hardware [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 925.993979] env[61867]: DEBUG nova.virt.hardware [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 925.994197] env[61867]: DEBUG nova.virt.hardware [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 925.994374] env[61867]: DEBUG nova.virt.hardware [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 925.994555] env[61867]: DEBUG nova.virt.hardware [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 925.995827] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef6b144c-4382-4d8c-8cd8-2f1e1fa48544 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.005208] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8177c048-1597-4694-88ca-95d072ed2d74 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.056932] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276876, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477517} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.057884] env[61867]: DEBUG nova.network.neutron [req-d512a7a3-cf9f-46be-bdbc-d37576859509 req-2d974286-e29c-4d60-bcb5-d3676c6e6037 service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 926.059691] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 6d20f463-9198-4590-8ec2-db471c64ba7c/6d20f463-9198-4590-8ec2-db471c64ba7c.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 926.059926] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 926.060568] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da8ad5fd-3226-4701-a2a8-3404d12ef358 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.069833] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 926.069833] env[61867]: value = "task-1276879" [ 926.069833] env[61867]: _type = "Task" [ 926.069833] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.077763] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276879, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.164731] env[61867]: DEBUG nova.network.neutron [req-d512a7a3-cf9f-46be-bdbc-d37576859509 req-2d974286-e29c-4d60-bcb5-d3676c6e6037 service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.192938] env[61867]: DEBUG oslo_vmware.api [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276877, 'name': ReconfigVM_Task, 'duration_secs': 0.206142} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.193288] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274394', 'volume_id': '9bcdabad-faa3-4b43-ac62-c59cb70802e4', 'name': 'volume-9bcdabad-faa3-4b43-ac62-c59cb70802e4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7479bf91-5aef-4e75-a127-7e82ae15a003', 'attached_at': '', 'detached_at': '', 'volume_id': '9bcdabad-faa3-4b43-ac62-c59cb70802e4', 'serial': '9bcdabad-faa3-4b43-ac62-c59cb70802e4'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 926.356444] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e6aed529-32af-431d-9685-a1d857dfa91c tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "5ec9eb1f-1d24-41d9-b5c7-a478a427c563" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.904s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.409975] env[61867]: DEBUG nova.network.neutron [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Successfully updated port: 211a1ec3-d618-4c1d-ac61-7e7b5681b9ed {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 926.590756] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276879, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070349} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.591310] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 926.592028] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9df06805-2fc6-4780-9e6d-0ae34cfefe34 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.599412] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca97aa43-98e8-4213-8c2b-6585c23eae7c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.617410] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d46af4b-e2b0-4807-abe8-27d097fc5804 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.629447] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 6d20f463-9198-4590-8ec2-db471c64ba7c/6d20f463-9198-4590-8ec2-db471c64ba7c.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 926.632318] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d708cd06-9ff6-4a72-a2b1-9bc104c74fa1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.676791] env[61867]: DEBUG oslo_concurrency.lockutils [req-d512a7a3-cf9f-46be-bdbc-d37576859509 req-2d974286-e29c-4d60-bcb5-d3676c6e6037 service nova] Releasing lock "refresh_cache-38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.679265] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94295e3f-9f6e-4ce5-9360-b6e5363b55f1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.682465] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 926.682465] env[61867]: value = "task-1276880" [ 926.682465] env[61867]: _type = "Task" [ 926.682465] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.690230] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67b7919-0fd5-48d7-bca5-f9419f4be7c4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.699348] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.712150] env[61867]: DEBUG nova.compute.provider_tree [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.813169] env[61867]: DEBUG nova.network.neutron [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Updating instance_info_cache with network_info: [{"id": "bad6d5ef-c6d0-448d-8a9b-01003e9872c7", "address": "fa:16:3e:45:db:29", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbad6d5ef-c6", "ovs_interfaceid": "bad6d5ef-c6d0-448d-8a9b-01003e9872c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.880306] env[61867]: DEBUG nova.network.neutron [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Successfully updated port: 017138c2-47c5-4033-a0b5-42707d74a494 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 926.915067] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "refresh_cache-8a7f4314-0fd4-49f9-8eb6-12baa0977a53" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.915339] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquired lock "refresh_cache-8a7f4314-0fd4-49f9-8eb6-12baa0977a53" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.915429] env[61867]: DEBUG nova.network.neutron [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 926.963918] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "refresh_cache-705a32b6-67f9-42cc-b4d0-f6d1783c68b5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.964180] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "refresh_cache-705a32b6-67f9-42cc-b4d0-f6d1783c68b5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.964382] env[61867]: DEBUG nova.network.neutron [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 927.011934] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "adc2732d-2a10-40ce-bb90-ed0762a36614" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.012237] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "adc2732d-2a10-40ce-bb90-ed0762a36614" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.012508] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "adc2732d-2a10-40ce-bb90-ed0762a36614-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.012890] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "adc2732d-2a10-40ce-bb90-ed0762a36614-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.013135] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "adc2732d-2a10-40ce-bb90-ed0762a36614-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.015262] env[61867]: INFO nova.compute.manager [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Terminating instance [ 927.017150] env[61867]: DEBUG nova.compute.manager [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 927.017359] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 927.020217] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c76ebe9-7c6d-46d1-8573-bf54df8ccb61 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.028018] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 927.028289] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-185b8b90-5e5d-4402-9071-d257ed09ad03 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.034689] env[61867]: DEBUG oslo_vmware.api [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 927.034689] env[61867]: value = "task-1276881" [ 927.034689] env[61867]: _type = "Task" [ 927.034689] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.043439] env[61867]: DEBUG oslo_vmware.api [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276881, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.197946] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276880, 'name': ReconfigVM_Task, 'duration_secs': 0.310389} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.198274] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 6d20f463-9198-4590-8ec2-db471c64ba7c/6d20f463-9198-4590-8ec2-db471c64ba7c.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.198938] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-35dc0bd5-b916-4cef-ba32-c4bd99666483 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.207262] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 927.207262] env[61867]: value = "task-1276882" [ 927.207262] env[61867]: _type = "Task" [ 927.207262] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.218825] env[61867]: DEBUG nova.scheduler.client.report [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.222740] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276882, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.263769] env[61867]: DEBUG nova.objects.instance [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lazy-loading 'flavor' on Instance uuid 7479bf91-5aef-4e75-a127-7e82ae15a003 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.317745] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "refresh_cache-9e1cbfa9-28e9-4bca-adfd-78bf25428106" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.493016] env[61867]: DEBUG nova.network.neutron [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 927.542061] env[61867]: DEBUG nova.compute.manager [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Received event network-vif-plugged-211a1ec3-d618-4c1d-ac61-7e7b5681b9ed {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.542389] env[61867]: DEBUG oslo_concurrency.lockutils [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] Acquiring lock "8a7f4314-0fd4-49f9-8eb6-12baa0977a53-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.542648] env[61867]: DEBUG oslo_concurrency.lockutils [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] Lock "8a7f4314-0fd4-49f9-8eb6-12baa0977a53-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.543552] env[61867]: DEBUG oslo_concurrency.lockutils [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] Lock "8a7f4314-0fd4-49f9-8eb6-12baa0977a53-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.543894] env[61867]: DEBUG nova.compute.manager [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] No waiting events found dispatching network-vif-plugged-211a1ec3-d618-4c1d-ac61-7e7b5681b9ed {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 927.544363] env[61867]: WARNING nova.compute.manager [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Received unexpected event network-vif-plugged-211a1ec3-d618-4c1d-ac61-7e7b5681b9ed for instance with vm_state building and task_state spawning. [ 927.544763] env[61867]: DEBUG nova.compute.manager [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Received event network-changed-211a1ec3-d618-4c1d-ac61-7e7b5681b9ed {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.545064] env[61867]: DEBUG nova.compute.manager [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Refreshing instance network info cache due to event network-changed-211a1ec3-d618-4c1d-ac61-7e7b5681b9ed. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 927.545398] env[61867]: DEBUG oslo_concurrency.lockutils [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] Acquiring lock "refresh_cache-8a7f4314-0fd4-49f9-8eb6-12baa0977a53" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.550288] env[61867]: DEBUG oslo_vmware.api [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276881, 'name': PowerOffVM_Task, 'duration_secs': 0.245122} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.550696] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 927.550955] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 927.551266] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1475673e-fe69-4699-9972-a054dd30e8a1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.663239] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 927.663742] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 927.664164] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleting the datastore file [datastore2] adc2732d-2a10-40ce-bb90-ed0762a36614 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.665817] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-85910ed7-6534-4557-92d8-74a10209f624 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.674198] env[61867]: DEBUG oslo_vmware.api [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 927.674198] env[61867]: value = "task-1276884" [ 927.674198] env[61867]: _type = "Task" [ 927.674198] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.686958] env[61867]: DEBUG oslo_vmware.api [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276884, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.718923] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276882, 'name': Rename_Task, 'duration_secs': 0.144142} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.721486] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 927.721486] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6bc4ae8f-3d57-4363-b3f2-5ffbcfbb59d1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.723748] env[61867]: DEBUG oslo_concurrency.lockutils [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.907s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.727946] env[61867]: DEBUG oslo_concurrency.lockutils [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.567s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.731664] env[61867]: INFO nova.compute.claims [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 927.733934] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 927.733934] env[61867]: value = "task-1276885" [ 927.733934] env[61867]: _type = "Task" [ 927.733934] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.746971] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276885, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.758844] env[61867]: INFO nova.scheduler.client.report [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Deleted allocations for instance ca2f6d18-f773-4875-83de-2f2be912c2f8 [ 927.768708] env[61867]: DEBUG oslo_concurrency.lockutils [None req-642973e2-a3e9-4169-9156-38b04ca40990 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "7479bf91-5aef-4e75-a127-7e82ae15a003" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.346s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.853788] env[61867]: DEBUG nova.network.neutron [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Updating instance_info_cache with network_info: [{"id": "211a1ec3-d618-4c1d-ac61-7e7b5681b9ed", "address": "fa:16:3e:b3:dd:4d", "network": {"id": "fc403caf-421f-4783-9807-5e61c4379a6d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-846804042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24180ce8aca142fb897e29d4853c20c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap211a1ec3-d6", "ovs_interfaceid": "211a1ec3-d618-4c1d-ac61-7e7b5681b9ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.875526] env[61867]: INFO nova.compute.manager [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Rescuing [ 927.875526] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.875526] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.875526] env[61867]: DEBUG nova.network.neutron [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 927.940570] env[61867]: DEBUG nova.network.neutron [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Updating instance_info_cache with network_info: [{"id": "254a9643-f941-493e-8c87-2932a6cc00ce", "address": "fa:16:3e:85:01:b4", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap254a9643-f9", "ovs_interfaceid": "254a9643-f941-493e-8c87-2932a6cc00ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.967954] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 927.968884] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec81ed6-658b-463e-a1f3-874eae9184e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.977839] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 927.978142] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fdf53b98-f4c4-4e11-9e05-af472493f21b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.043052] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 928.043251] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 928.043368] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleting the datastore file [datastore1] 9e1cbfa9-28e9-4bca-adfd-78bf25428106 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 928.043672] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a702df2c-7d16-4a9b-9046-fe2e0c5cb3ec {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.051208] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 928.051208] env[61867]: value = "task-1276887" [ 928.051208] env[61867]: _type = "Task" [ 928.051208] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.060471] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276887, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.182272] env[61867]: DEBUG oslo_vmware.api [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276884, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.189756} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.182553] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 928.182776] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 928.182965] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 928.183166] env[61867]: INFO nova.compute.manager [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Took 1.17 seconds to destroy the instance on the hypervisor. [ 928.183421] env[61867]: DEBUG oslo.service.loopingcall [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.183622] env[61867]: DEBUG nova.compute.manager [-] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 928.183744] env[61867]: DEBUG nova.network.neutron [-] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 928.247589] env[61867]: DEBUG oslo_vmware.api [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276885, 'name': PowerOnVM_Task, 'duration_secs': 0.466842} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.247589] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 928.248024] env[61867]: INFO nova.compute.manager [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Took 7.07 seconds to spawn the instance on the hypervisor. [ 928.248398] env[61867]: DEBUG nova.compute.manager [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 928.250023] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-675444af-886b-4772-bb0d-9e1c79e6f7b3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.271025] env[61867]: DEBUG oslo_concurrency.lockutils [None req-49cd35b9-1201-4fe4-86bf-7c7be76af694 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "ca2f6d18-f773-4875-83de-2f2be912c2f8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.965s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.355466] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Releasing lock "refresh_cache-8a7f4314-0fd4-49f9-8eb6-12baa0977a53" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.355965] env[61867]: DEBUG nova.compute.manager [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Instance network_info: |[{"id": "211a1ec3-d618-4c1d-ac61-7e7b5681b9ed", "address": "fa:16:3e:b3:dd:4d", "network": {"id": "fc403caf-421f-4783-9807-5e61c4379a6d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-846804042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24180ce8aca142fb897e29d4853c20c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap211a1ec3-d6", "ovs_interfaceid": "211a1ec3-d618-4c1d-ac61-7e7b5681b9ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 928.357217] env[61867]: DEBUG oslo_concurrency.lockutils [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] Acquired lock "refresh_cache-8a7f4314-0fd4-49f9-8eb6-12baa0977a53" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.357577] env[61867]: DEBUG nova.network.neutron [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Refreshing network info cache for port 211a1ec3-d618-4c1d-ac61-7e7b5681b9ed {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 928.359771] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:dd:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '47ca1ce6-8148-48d5-bcfe-89e39b73914e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '211a1ec3-d618-4c1d-ac61-7e7b5681b9ed', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 928.373590] env[61867]: DEBUG oslo.service.loopingcall [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.379984] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 928.380925] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-83d593d2-20cd-4682-87cb-b609a6ff1cf7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.419514] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 928.419514] env[61867]: value = "task-1276888" [ 928.419514] env[61867]: _type = "Task" [ 928.419514] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.431622] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276888, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.444121] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "refresh_cache-705a32b6-67f9-42cc-b4d0-f6d1783c68b5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.566378] env[61867]: DEBUG oslo_vmware.api [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276887, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.252247} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.566378] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 928.566771] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 928.566868] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 928.622444] env[61867]: INFO nova.scheduler.client.report [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleted allocations for instance 9e1cbfa9-28e9-4bca-adfd-78bf25428106 [ 928.774237] env[61867]: INFO nova.compute.manager [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Took 12.44 seconds to build instance. [ 928.811500] env[61867]: DEBUG nova.network.neutron [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Updated VIF entry in instance network info cache for port 211a1ec3-d618-4c1d-ac61-7e7b5681b9ed. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 928.812336] env[61867]: DEBUG nova.network.neutron [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Updating instance_info_cache with network_info: [{"id": "211a1ec3-d618-4c1d-ac61-7e7b5681b9ed", "address": "fa:16:3e:b3:dd:4d", "network": {"id": "fc403caf-421f-4783-9807-5e61c4379a6d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-846804042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24180ce8aca142fb897e29d4853c20c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap211a1ec3-d6", "ovs_interfaceid": "211a1ec3-d618-4c1d-ac61-7e7b5681b9ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.866861] env[61867]: DEBUG nova.network.neutron [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Updating instance_info_cache with network_info: [{"id": "4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1", "address": "fa:16:3e:ce:d6:b5", "network": {"id": "ab6ec5b9-8193-43c9-a3ce-d63f31e5a001", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1573384141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35aa85105b394369b17ef4ce51a49066", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f63bd0e-c1", "ovs_interfaceid": "4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.931132] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276888, 'name': CreateVM_Task, 'duration_secs': 0.356502} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.931577] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 928.934741] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.935096] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.935527] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 928.936098] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd741e46-67da-4b72-86ad-290fb2e8a06f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.943155] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 928.943155] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5231012f-611c-29f8-880f-5cdff4481936" [ 928.943155] env[61867]: _type = "Task" [ 928.943155] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.957181] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5231012f-611c-29f8-880f-5cdff4481936, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.978987] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a1cc4ea-2d1d-4d6c-9354-4ae1b5e6ee8d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.983496] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920ebed6-4dbd-4d95-87a7-08bda8d27611 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.007886] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d4ebc3a-4157-4ed4-878b-2650c27116ce {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.013870] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc09745d-5f3c-446a-96dd-2544212f9f77 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.021433] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Updating instance '705a32b6-67f9-42cc-b4d0-f6d1783c68b5' progress to 83 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 929.050013] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07529134-16e3-4e6f-baec-7cde5893230f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.061510] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcfb99d-cf17-4e6c-b6b7-4e937c9d7dc5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.081288] env[61867]: DEBUG nova.compute.provider_tree [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.086325] env[61867]: DEBUG nova.network.neutron [-] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.128441] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.249464] env[61867]: DEBUG nova.compute.manager [req-dc4659ba-85d8-4179-b147-09936981f49b req-67438b91-bf86-479d-b8c8-c7cc33d2e7fe service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Received event network-vif-plugged-74ad152e-530d-49fd-bdc1-30e366f5f9ac {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.249773] env[61867]: DEBUG oslo_concurrency.lockutils [req-dc4659ba-85d8-4179-b147-09936981f49b req-67438b91-bf86-479d-b8c8-c7cc33d2e7fe service nova] Acquiring lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.250040] env[61867]: DEBUG oslo_concurrency.lockutils [req-dc4659ba-85d8-4179-b147-09936981f49b req-67438b91-bf86-479d-b8c8-c7cc33d2e7fe service nova] Lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.250258] env[61867]: DEBUG oslo_concurrency.lockutils [req-dc4659ba-85d8-4179-b147-09936981f49b req-67438b91-bf86-479d-b8c8-c7cc33d2e7fe service nova] Lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.250517] env[61867]: DEBUG nova.compute.manager [req-dc4659ba-85d8-4179-b147-09936981f49b req-67438b91-bf86-479d-b8c8-c7cc33d2e7fe service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] No waiting events found dispatching network-vif-plugged-74ad152e-530d-49fd-bdc1-30e366f5f9ac {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 929.250606] env[61867]: WARNING nova.compute.manager [req-dc4659ba-85d8-4179-b147-09936981f49b req-67438b91-bf86-479d-b8c8-c7cc33d2e7fe service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Received unexpected event network-vif-plugged-74ad152e-530d-49fd-bdc1-30e366f5f9ac for instance with vm_state building and task_state spawning. [ 929.276757] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0aac84a7-5a12-4c96-924e-0a4a3cd4d06a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "6d20f463-9198-4590-8ec2-db471c64ba7c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.971s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.313997] env[61867]: DEBUG nova.network.neutron [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Successfully updated port: 74ad152e-530d-49fd-bdc1-30e366f5f9ac {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 929.318261] env[61867]: DEBUG oslo_concurrency.lockutils [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] Releasing lock "refresh_cache-8a7f4314-0fd4-49f9-8eb6-12baa0977a53" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.318261] env[61867]: DEBUG nova.compute.manager [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Received event network-vif-plugged-017138c2-47c5-4033-a0b5-42707d74a494 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.318261] env[61867]: DEBUG oslo_concurrency.lockutils [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] Acquiring lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.318261] env[61867]: DEBUG oslo_concurrency.lockutils [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] Lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.318261] env[61867]: DEBUG oslo_concurrency.lockutils [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] Lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.318261] env[61867]: DEBUG nova.compute.manager [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] No waiting events found dispatching network-vif-plugged-017138c2-47c5-4033-a0b5-42707d74a494 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 929.318261] env[61867]: WARNING nova.compute.manager [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Received unexpected event network-vif-plugged-017138c2-47c5-4033-a0b5-42707d74a494 for instance with vm_state building and task_state spawning. [ 929.318261] env[61867]: DEBUG nova.compute.manager [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Received event network-changed-017138c2-47c5-4033-a0b5-42707d74a494 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.318261] env[61867]: DEBUG nova.compute.manager [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Refreshing instance network info cache due to event network-changed-017138c2-47c5-4033-a0b5-42707d74a494. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 929.318261] env[61867]: DEBUG oslo_concurrency.lockutils [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] Acquiring lock "refresh_cache-38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.318261] env[61867]: DEBUG oslo_concurrency.lockutils [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] Acquired lock "refresh_cache-38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.319555] env[61867]: DEBUG nova.network.neutron [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Refreshing network info cache for port 017138c2-47c5-4033-a0b5-42707d74a494 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 929.371184] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Releasing lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.455359] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5231012f-611c-29f8-880f-5cdff4481936, 'name': SearchDatastore_Task, 'duration_secs': 0.011311} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.455696] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.456176] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 929.456491] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.456702] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.457163] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 929.457242] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1a1e7e1b-dffc-4639-b884-111cb5049678 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.466607] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 929.466813] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 929.467546] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-562dca2e-286d-4681-9bb1-033c54dc7395 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.478235] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 929.478235] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5243d00a-f354-e487-1c0f-5c4738ce9e8b" [ 929.478235] env[61867]: _type = "Task" [ 929.478235] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.491129] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5243d00a-f354-e487-1c0f-5c4738ce9e8b, 'name': SearchDatastore_Task, 'duration_secs': 0.009536} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.491490] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57e3e43c-34de-42c8-bffa-ae0f346e7a2f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.502453] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 929.502453] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ea24c4-f355-d660-c6d3-77d023cb20f1" [ 929.502453] env[61867]: _type = "Task" [ 929.502453] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.520049] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ea24c4-f355-d660-c6d3-77d023cb20f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.553364] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 929.553364] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b97c9e92-848f-42cb-a9e3-bb147f657355 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.564820] env[61867]: DEBUG oslo_vmware.api [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 929.564820] env[61867]: value = "task-1276889" [ 929.564820] env[61867]: _type = "Task" [ 929.564820] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.572871] env[61867]: DEBUG oslo_vmware.api [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276889, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.587568] env[61867]: DEBUG nova.scheduler.client.report [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 929.589762] env[61867]: INFO nova.compute.manager [-] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Took 1.41 seconds to deallocate network for instance. [ 929.733503] env[61867]: DEBUG nova.compute.manager [req-f6499cd1-79e4-4ab4-b344-0c710686ad94 req-7fe0d3c6-4931-47e9-8f16-631b3a596e12 service nova] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Received event network-vif-unplugged-bad6d5ef-c6d0-448d-8a9b-01003e9872c7 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.733629] env[61867]: DEBUG oslo_concurrency.lockutils [req-f6499cd1-79e4-4ab4-b344-0c710686ad94 req-7fe0d3c6-4931-47e9-8f16-631b3a596e12 service nova] Acquiring lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.734700] env[61867]: DEBUG oslo_concurrency.lockutils [req-f6499cd1-79e4-4ab4-b344-0c710686ad94 req-7fe0d3c6-4931-47e9-8f16-631b3a596e12 service nova] Lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.735062] env[61867]: DEBUG oslo_concurrency.lockutils [req-f6499cd1-79e4-4ab4-b344-0c710686ad94 req-7fe0d3c6-4931-47e9-8f16-631b3a596e12 service nova] Lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.735062] env[61867]: DEBUG nova.compute.manager [req-f6499cd1-79e4-4ab4-b344-0c710686ad94 req-7fe0d3c6-4931-47e9-8f16-631b3a596e12 service nova] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] No waiting events found dispatching network-vif-unplugged-bad6d5ef-c6d0-448d-8a9b-01003e9872c7 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 929.735163] env[61867]: DEBUG nova.compute.manager [req-f6499cd1-79e4-4ab4-b344-0c710686ad94 req-7fe0d3c6-4931-47e9-8f16-631b3a596e12 service nova] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Received event network-vif-unplugged-bad6d5ef-c6d0-448d-8a9b-01003e9872c7 for instance with task_state deleting. {{(pid=61867) _process_instance_event /opt/stack/nova/nova/compute/manager.py:10909}} [ 929.735536] env[61867]: DEBUG nova.compute.manager [req-f6499cd1-79e4-4ab4-b344-0c710686ad94 req-7fe0d3c6-4931-47e9-8f16-631b3a596e12 service nova] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Received event network-changed-bad6d5ef-c6d0-448d-8a9b-01003e9872c7 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.735779] env[61867]: DEBUG nova.compute.manager [req-f6499cd1-79e4-4ab4-b344-0c710686ad94 req-7fe0d3c6-4931-47e9-8f16-631b3a596e12 service nova] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Refreshing instance network info cache due to event network-changed-bad6d5ef-c6d0-448d-8a9b-01003e9872c7. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 929.735985] env[61867]: DEBUG oslo_concurrency.lockutils [req-f6499cd1-79e4-4ab4-b344-0c710686ad94 req-7fe0d3c6-4931-47e9-8f16-631b3a596e12 service nova] Acquiring lock "refresh_cache-9e1cbfa9-28e9-4bca-adfd-78bf25428106" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.736144] env[61867]: DEBUG oslo_concurrency.lockutils [req-f6499cd1-79e4-4ab4-b344-0c710686ad94 req-7fe0d3c6-4931-47e9-8f16-631b3a596e12 service nova] Acquired lock "refresh_cache-9e1cbfa9-28e9-4bca-adfd-78bf25428106" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.736563] env[61867]: DEBUG nova.network.neutron [req-f6499cd1-79e4-4ab4-b344-0c710686ad94 req-7fe0d3c6-4931-47e9-8f16-631b3a596e12 service nova] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Refreshing network info cache for port bad6d5ef-c6d0-448d-8a9b-01003e9872c7 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 929.822643] env[61867]: DEBUG oslo_concurrency.lockutils [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquiring lock "refresh_cache-38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.884711] env[61867]: DEBUG nova.network.neutron [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 929.904516] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 929.904878] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-18e0ed3e-7a61-4ab7-ba8d-ea66e316da02 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.913080] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 929.913080] env[61867]: value = "task-1276890" [ 929.913080] env[61867]: _type = "Task" [ 929.913080] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.923378] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276890, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.016109] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ea24c4-f355-d660-c6d3-77d023cb20f1, 'name': SearchDatastore_Task, 'duration_secs': 0.014323} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.016408] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.016658] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 8a7f4314-0fd4-49f9-8eb6-12baa0977a53/8a7f4314-0fd4-49f9-8eb6-12baa0977a53.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 930.017279] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-99790b71-1712-4f76-a76d-a8510fd45711 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.027215] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 930.027215] env[61867]: value = "task-1276891" [ 930.027215] env[61867]: _type = "Task" [ 930.027215] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.038599] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276891, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.079026] env[61867]: DEBUG oslo_vmware.api [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276889, 'name': PowerOnVM_Task, 'duration_secs': 0.508767} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.079026] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 930.079026] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2949d2bd-7dd9-4269-968b-3a62c6136c9f tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Updating instance '705a32b6-67f9-42cc-b4d0-f6d1783c68b5' progress to 100 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 930.094860] env[61867]: DEBUG oslo_concurrency.lockutils [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.095521] env[61867]: DEBUG nova.compute.manager [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 930.099036] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.111s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.103034] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.128318] env[61867]: DEBUG nova.network.neutron [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.205530] env[61867]: DEBUG oslo_concurrency.lockutils [None req-633e67b6-a50b-4c01-af8d-8044424e040e tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.425514] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276890, 'name': PowerOffVM_Task, 'duration_secs': 0.282264} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.425913] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 930.426770] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebc6e3b8-92e4-4617-bb7a-e0f2fa825ced {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.449793] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e674b5a7-2511-4741-b067-0ff00e4fadac {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.486656] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 930.487065] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e337f7eb-6bc3-44cf-8727-9cd15fa0ff83 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.503061] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 930.503061] env[61867]: value = "task-1276892" [ 930.503061] env[61867]: _type = "Task" [ 930.503061] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.516357] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] VM already powered off {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 930.516600] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 930.516860] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.517025] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.517217] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 930.517482] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a2efb78-7e96-4910-aebd-45a81d826612 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.527135] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 930.527210] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 930.527949] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c87c3aa5-f6b2-4973-9ff0-91946bf7bcfc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.539361] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276891, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508308} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.542735] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 8a7f4314-0fd4-49f9-8eb6-12baa0977a53/8a7f4314-0fd4-49f9-8eb6-12baa0977a53.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 930.542984] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 930.543324] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 930.543324] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5253f0f8-9d55-e20f-9bba-9c888e9296eb" [ 930.543324] env[61867]: _type = "Task" [ 930.543324] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.543525] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9f5f6607-ed32-4fe6-86d8-259ec8f1c090 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.555830] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5253f0f8-9d55-e20f-9bba-9c888e9296eb, 'name': SearchDatastore_Task, 'duration_secs': 0.010634} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.557820] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 930.557820] env[61867]: value = "task-1276893" [ 930.557820] env[61867]: _type = "Task" [ 930.557820] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.558385] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e21642ff-5aef-4335-8c62-9b3eacb10618 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.569831] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276893, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.570198] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 930.570198] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5296f63c-8b74-bab5-3b69-1adcfc6ecd3f" [ 930.570198] env[61867]: _type = "Task" [ 930.570198] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.585895] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5296f63c-8b74-bab5-3b69-1adcfc6ecd3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.602889] env[61867]: DEBUG nova.compute.utils [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 930.604806] env[61867]: DEBUG nova.compute.manager [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 930.604979] env[61867]: DEBUG nova.network.neutron [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 930.609273] env[61867]: INFO nova.compute.claims [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.632481] env[61867]: DEBUG oslo_concurrency.lockutils [req-349c3069-c5ad-4865-af1e-3a64e10feb6a req-200ea017-7367-4ae6-8ee5-665519267fdf service nova] Releasing lock "refresh_cache-38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.632481] env[61867]: DEBUG oslo_concurrency.lockutils [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquired lock "refresh_cache-38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.632481] env[61867]: DEBUG nova.network.neutron [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 930.660814] env[61867]: DEBUG nova.policy [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '86b7e5f767b94dc8a88b1e7f69c923e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '05dd2be830f9410fbb90415cc13ff6bf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 930.942153] env[61867]: DEBUG nova.network.neutron [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Successfully created port: 1a175a33-e511-4f6a-a971-708994d51259 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 930.945041] env[61867]: DEBUG nova.network.neutron [req-f6499cd1-79e4-4ab4-b344-0c710686ad94 req-7fe0d3c6-4931-47e9-8f16-631b3a596e12 service nova] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Updated VIF entry in instance network info cache for port bad6d5ef-c6d0-448d-8a9b-01003e9872c7. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 930.945409] env[61867]: DEBUG nova.network.neutron [req-f6499cd1-79e4-4ab4-b344-0c710686ad94 req-7fe0d3c6-4931-47e9-8f16-631b3a596e12 service nova] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Updating instance_info_cache with network_info: [{"id": "bad6d5ef-c6d0-448d-8a9b-01003e9872c7", "address": "fa:16:3e:45:db:29", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": null, "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapbad6d5ef-c6", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.075029] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276893, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079866} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.082363] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 931.083268] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-addaaa77-11db-45ab-bcd7-db08f2c8b05f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.097539] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5296f63c-8b74-bab5-3b69-1adcfc6ecd3f, 'name': SearchDatastore_Task, 'duration_secs': 0.011152} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.097950] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.099730] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 7479bf91-5aef-4e75-a127-7e82ae15a003/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk. {{(pid=61867) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 931.099730] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e0204361-0f3a-4a97-80ce-6fca8c310a3a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.121548] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 8a7f4314-0fd4-49f9-8eb6-12baa0977a53/8a7f4314-0fd4-49f9-8eb6-12baa0977a53.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.121548] env[61867]: DEBUG nova.compute.manager [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 931.125209] env[61867]: INFO nova.compute.resource_tracker [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updating resource usage from migration ca3d0401-216e-4d3c-a441-9d1297b25042 [ 931.128213] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d462b07e-5c64-4cc3-a075-754d2a6b425a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.146502] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 931.146502] env[61867]: value = "task-1276894" [ 931.146502] env[61867]: _type = "Task" [ 931.146502] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.155956] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 931.155956] env[61867]: value = "task-1276895" [ 931.155956] env[61867]: _type = "Task" [ 931.155956] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.161441] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276894, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.173539] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276895, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.222293] env[61867]: DEBUG nova.network.neutron [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 931.282138] env[61867]: DEBUG oslo_concurrency.lockutils [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "4d7bfb31-d565-49e4-8c51-0122acebba2d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.282410] env[61867]: DEBUG oslo_concurrency.lockutils [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "4d7bfb31-d565-49e4-8c51-0122acebba2d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.435300] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb2a9fac-b018-4f0b-a447-b1c9a3fcfde8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.446124] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5be2e92-37b8-47d6-a4fe-7231dbcd0a6c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.451250] env[61867]: DEBUG oslo_concurrency.lockutils [req-f6499cd1-79e4-4ab4-b344-0c710686ad94 req-7fe0d3c6-4931-47e9-8f16-631b3a596e12 service nova] Releasing lock "refresh_cache-9e1cbfa9-28e9-4bca-adfd-78bf25428106" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.451619] env[61867]: DEBUG nova.compute.manager [req-f6499cd1-79e4-4ab4-b344-0c710686ad94 req-7fe0d3c6-4931-47e9-8f16-631b3a596e12 service nova] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Received event network-vif-deleted-a101c251-3c8e-4c0f-8528-aba0e8002f42 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.494832] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abfba451-66b7-4db5-ba03-a1832776dcba {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.507401] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66249fbe-062b-4ed7-8018-c333c7048485 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.529712] env[61867]: DEBUG nova.compute.provider_tree [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.662128] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276894, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.674189] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276895, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.746265] env[61867]: DEBUG nova.compute.manager [req-15d63e63-71e4-421e-8d12-de5c4e4ade0a req-ce45eb6c-2495-4c6f-8a0f-2baa2957c6a4 service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Received event network-changed-74ad152e-530d-49fd-bdc1-30e366f5f9ac {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.746498] env[61867]: DEBUG nova.compute.manager [req-15d63e63-71e4-421e-8d12-de5c4e4ade0a req-ce45eb6c-2495-4c6f-8a0f-2baa2957c6a4 service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Refreshing instance network info cache due to event network-changed-74ad152e-530d-49fd-bdc1-30e366f5f9ac. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 931.746798] env[61867]: DEBUG oslo_concurrency.lockutils [req-15d63e63-71e4-421e-8d12-de5c4e4ade0a req-ce45eb6c-2495-4c6f-8a0f-2baa2957c6a4 service nova] Acquiring lock "refresh_cache-38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.784974] env[61867]: DEBUG nova.compute.manager [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 932.033113] env[61867]: DEBUG nova.scheduler.client.report [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 932.155483] env[61867]: DEBUG nova.compute.manager [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 932.163801] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276894, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.81953} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.169306] env[61867]: INFO nova.virt.vmwareapi.ds_util [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 7479bf91-5aef-4e75-a127-7e82ae15a003/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk. [ 932.170143] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0233ee9d-6991-4441-b0aa-cb24787ff1fc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.180870] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276895, 'name': ReconfigVM_Task, 'duration_secs': 0.756618} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.183067] env[61867]: DEBUG nova.virt.hardware [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 932.183304] env[61867]: DEBUG nova.virt.hardware [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 932.183466] env[61867]: DEBUG nova.virt.hardware [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 932.183653] env[61867]: DEBUG nova.virt.hardware [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 932.183876] env[61867]: DEBUG nova.virt.hardware [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 932.184069] env[61867]: DEBUG nova.virt.hardware [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 932.184290] env[61867]: DEBUG nova.virt.hardware [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 932.184453] env[61867]: DEBUG nova.virt.hardware [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 932.184622] env[61867]: DEBUG nova.virt.hardware [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 932.184784] env[61867]: DEBUG nova.virt.hardware [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 932.184959] env[61867]: DEBUG nova.virt.hardware [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 932.202158] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 8a7f4314-0fd4-49f9-8eb6-12baa0977a53/8a7f4314-0fd4-49f9-8eb6-12baa0977a53.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.203656] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4dc7a1d-d280-43b0-b07b-71eb41b1f81b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.213522] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 7479bf91-5aef-4e75-a127-7e82ae15a003/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 932.213804] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-de440197-7631-460a-baff-723dc65d173b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.215796] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97fb136b-3b75-45ad-93bf-0d1c9eb414b2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.238204] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e79dab-d829-43bd-9c93-ce5d3c984a10 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.242474] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 932.242474] env[61867]: value = "task-1276896" [ 932.242474] env[61867]: _type = "Task" [ 932.242474] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.242758] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 932.242758] env[61867]: value = "task-1276897" [ 932.242758] env[61867]: _type = "Task" [ 932.242758] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.259444] env[61867]: DEBUG nova.network.neutron [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Updating instance_info_cache with network_info: [{"id": "d9630611-bb16-428c-ade6-78ff0355d81d", "address": "fa:16:3e:b0:dd:b6", "network": {"id": "b49f45d7-ec84-4a83-b8af-9aac5a49ce7a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1342995777", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.73", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9630611-bb", "ovs_interfaceid": "d9630611-bb16-428c-ade6-78ff0355d81d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "017138c2-47c5-4033-a0b5-42707d74a494", "address": "fa:16:3e:b6:0f:c7", "network": {"id": "fbd6b941-c8f4-4f3a-89c9-d424d9c30db3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-735604480", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.197", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap017138c2-47", "ovs_interfaceid": "017138c2-47c5-4033-a0b5-42707d74a494", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "74ad152e-530d-49fd-bdc1-30e366f5f9ac", "address": "fa:16:3e:3d:21:12", "network": {"id": "b49f45d7-ec84-4a83-b8af-9aac5a49ce7a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1342995777", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.148", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74ad152e-53", "ovs_interfaceid": "74ad152e-530d-49fd-bdc1-30e366f5f9ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.272897] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276897, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.275922] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276896, 'name': Rename_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.299956] env[61867]: DEBUG oslo_concurrency.lockutils [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.300299] env[61867]: DEBUG oslo_concurrency.lockutils [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.300544] env[61867]: DEBUG nova.compute.manager [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Going to confirm migration 2 {{(pid=61867) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 932.311359] env[61867]: DEBUG oslo_concurrency.lockutils [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.526282] env[61867]: DEBUG nova.network.neutron [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Successfully updated port: 1a175a33-e511-4f6a-a971-708994d51259 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 932.540017] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.439s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.540017] env[61867]: INFO nova.compute.manager [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Migrating [ 932.545588] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.417s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.546200] env[61867]: DEBUG nova.objects.instance [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lazy-loading 'resources' on Instance uuid 9e1cbfa9-28e9-4bca-adfd-78bf25428106 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.765740] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276896, 'name': Rename_Task, 'duration_secs': 0.229777} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.766945] env[61867]: DEBUG oslo_concurrency.lockutils [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Releasing lock "refresh_cache-38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.766945] env[61867]: DEBUG nova.compute.manager [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Instance network_info: |[{"id": "d9630611-bb16-428c-ade6-78ff0355d81d", "address": "fa:16:3e:b0:dd:b6", "network": {"id": "b49f45d7-ec84-4a83-b8af-9aac5a49ce7a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1342995777", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.73", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9630611-bb", "ovs_interfaceid": "d9630611-bb16-428c-ade6-78ff0355d81d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "017138c2-47c5-4033-a0b5-42707d74a494", "address": "fa:16:3e:b6:0f:c7", "network": {"id": "fbd6b941-c8f4-4f3a-89c9-d424d9c30db3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-735604480", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.197", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap017138c2-47", "ovs_interfaceid": "017138c2-47c5-4033-a0b5-42707d74a494", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "74ad152e-530d-49fd-bdc1-30e366f5f9ac", "address": "fa:16:3e:3d:21:12", "network": {"id": "b49f45d7-ec84-4a83-b8af-9aac5a49ce7a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1342995777", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.148", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74ad152e-53", "ovs_interfaceid": "74ad152e-530d-49fd-bdc1-30e366f5f9ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 932.767317] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276897, 'name': ReconfigVM_Task, 'duration_secs': 0.349094} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.767699] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 932.768113] env[61867]: DEBUG oslo_concurrency.lockutils [req-15d63e63-71e4-421e-8d12-de5c4e4ade0a req-ce45eb6c-2495-4c6f-8a0f-2baa2957c6a4 service nova] Acquired lock "refresh_cache-38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.768416] env[61867]: DEBUG nova.network.neutron [req-15d63e63-71e4-421e-8d12-de5c4e4ade0a req-ce45eb6c-2495-4c6f-8a0f-2baa2957c6a4 service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Refreshing network info cache for port 74ad152e-530d-49fd-bdc1-30e366f5f9ac {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 932.770031] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:dd:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c883fb98-d172-4510-8cf4-07aafdf771af', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd9630611-bb16-428c-ade6-78ff0355d81d', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:0f:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7894814c-6be3-4b80-a08e-4a771bc05dd1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '017138c2-47c5-4033-a0b5-42707d74a494', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:21:12', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c883fb98-d172-4510-8cf4-07aafdf771af', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '74ad152e-530d-49fd-bdc1-30e366f5f9ac', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 932.784740] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Creating folder: Project (60b35760e3e14245aea3600d36c838dc). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 932.785844] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 7479bf91-5aef-4e75-a127-7e82ae15a003/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.790021] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-189fccc4-d0cf-400c-9026-f0fe87dfd0e1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.790988] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ca5a8c52-af50-4f1d-83cd-669971c7954b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.793360] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a0149f-8f0d-45ba-b284-5f651f28dea7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.796481] env[61867]: DEBUG nova.compute.manager [req-7709fb9a-e91f-49f7-8742-3a7b37a62162 req-f4dff2a7-b3f9-409f-ab95-d61e8e4b4a36 service nova] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Received event network-vif-plugged-1a175a33-e511-4f6a-a971-708994d51259 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.796826] env[61867]: DEBUG oslo_concurrency.lockutils [req-7709fb9a-e91f-49f7-8742-3a7b37a62162 req-f4dff2a7-b3f9-409f-ab95-d61e8e4b4a36 service nova] Acquiring lock "3905193e-04da-439a-bf6c-16f638a692bf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.796940] env[61867]: DEBUG oslo_concurrency.lockutils [req-7709fb9a-e91f-49f7-8742-3a7b37a62162 req-f4dff2a7-b3f9-409f-ab95-d61e8e4b4a36 service nova] Lock "3905193e-04da-439a-bf6c-16f638a692bf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.797079] env[61867]: DEBUG oslo_concurrency.lockutils [req-7709fb9a-e91f-49f7-8742-3a7b37a62162 req-f4dff2a7-b3f9-409f-ab95-d61e8e4b4a36 service nova] Lock "3905193e-04da-439a-bf6c-16f638a692bf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.797255] env[61867]: DEBUG nova.compute.manager [req-7709fb9a-e91f-49f7-8742-3a7b37a62162 req-f4dff2a7-b3f9-409f-ab95-d61e8e4b4a36 service nova] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] No waiting events found dispatching network-vif-plugged-1a175a33-e511-4f6a-a971-708994d51259 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 932.797418] env[61867]: WARNING nova.compute.manager [req-7709fb9a-e91f-49f7-8742-3a7b37a62162 req-f4dff2a7-b3f9-409f-ab95-d61e8e4b4a36 service nova] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Received unexpected event network-vif-plugged-1a175a33-e511-4f6a-a971-708994d51259 for instance with vm_state building and task_state spawning. [ 932.835507] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7392a405-6f6c-4682-a038-34c9e8b01192 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.845921] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 932.845921] env[61867]: value = "task-1276899" [ 932.845921] env[61867]: _type = "Task" [ 932.845921] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.846711] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Created folder: Project (60b35760e3e14245aea3600d36c838dc) in parent group-v274258. [ 932.846882] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Creating folder: Instances. Parent ref: group-v274397. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 932.849444] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-894ea2e3-f2ae-476a-8e5b-1ee6d326bc4f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.857941] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 932.857941] env[61867]: value = "task-1276900" [ 932.857941] env[61867]: _type = "Task" [ 932.857941] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.861721] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276899, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.866384] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Created folder: Instances in parent group-v274397. [ 932.866635] env[61867]: DEBUG oslo.service.loopingcall [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.867669] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 932.867669] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a38ca676-dca6-49a8-a95b-11c8fd4ae4e9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.890193] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276900, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.897103] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 932.897103] env[61867]: value = "task-1276902" [ 932.897103] env[61867]: _type = "Task" [ 932.897103] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.907018] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276902, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.914511] env[61867]: DEBUG oslo_concurrency.lockutils [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "refresh_cache-705a32b6-67f9-42cc-b4d0-f6d1783c68b5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.914818] env[61867]: DEBUG oslo_concurrency.lockutils [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "refresh_cache-705a32b6-67f9-42cc-b4d0-f6d1783c68b5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.914956] env[61867]: DEBUG nova.network.neutron [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 932.915151] env[61867]: DEBUG nova.objects.instance [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lazy-loading 'info_cache' on Instance uuid 705a32b6-67f9-42cc-b4d0-f6d1783c68b5 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.029369] env[61867]: DEBUG oslo_concurrency.lockutils [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "refresh_cache-3905193e-04da-439a-bf6c-16f638a692bf" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.029560] env[61867]: DEBUG oslo_concurrency.lockutils [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquired lock "refresh_cache-3905193e-04da-439a-bf6c-16f638a692bf" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.030575] env[61867]: DEBUG nova.network.neutron [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 933.060019] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.060019] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.060019] env[61867]: DEBUG nova.network.neutron [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 933.060019] env[61867]: DEBUG nova.objects.instance [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lazy-loading 'numa_topology' on Instance uuid 9e1cbfa9-28e9-4bca-adfd-78bf25428106 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.061584] env[61867]: DEBUG nova.network.neutron [req-15d63e63-71e4-421e-8d12-de5c4e4ade0a req-ce45eb6c-2495-4c6f-8a0f-2baa2957c6a4 service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Updated VIF entry in instance network info cache for port 74ad152e-530d-49fd-bdc1-30e366f5f9ac. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 933.062163] env[61867]: DEBUG nova.network.neutron [req-15d63e63-71e4-421e-8d12-de5c4e4ade0a req-ce45eb6c-2495-4c6f-8a0f-2baa2957c6a4 service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Updating instance_info_cache with network_info: [{"id": "d9630611-bb16-428c-ade6-78ff0355d81d", "address": "fa:16:3e:b0:dd:b6", "network": {"id": "b49f45d7-ec84-4a83-b8af-9aac5a49ce7a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1342995777", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.73", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9630611-bb", "ovs_interfaceid": "d9630611-bb16-428c-ade6-78ff0355d81d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "017138c2-47c5-4033-a0b5-42707d74a494", "address": "fa:16:3e:b6:0f:c7", "network": {"id": "fbd6b941-c8f4-4f3a-89c9-d424d9c30db3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-735604480", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.197", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7894814c-6be3-4b80-a08e-4a771bc05dd1", "external-id": "nsx-vlan-transportzone-948", "segmentation_id": 948, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap017138c2-47", "ovs_interfaceid": "017138c2-47c5-4033-a0b5-42707d74a494", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "74ad152e-530d-49fd-bdc1-30e366f5f9ac", "address": "fa:16:3e:3d:21:12", "network": {"id": "b49f45d7-ec84-4a83-b8af-9aac5a49ce7a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1342995777", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.148", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74ad152e-53", "ovs_interfaceid": "74ad152e-530d-49fd-bdc1-30e366f5f9ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.362437] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276899, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.371329] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276900, 'name': ReconfigVM_Task, 'duration_secs': 0.247379} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.371831] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 933.373994] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a792e21-6ece-44fb-a25e-50ef00762576 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.380039] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 933.380039] env[61867]: value = "task-1276903" [ 933.380039] env[61867]: _type = "Task" [ 933.380039] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.387819] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276903, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.407590] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276902, 'name': CreateVM_Task, 'duration_secs': 0.477665} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.407590] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 933.410018] env[61867]: DEBUG oslo_concurrency.lockutils [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.410018] env[61867]: DEBUG oslo_concurrency.lockutils [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.410018] env[61867]: DEBUG oslo_concurrency.lockutils [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 933.410018] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-183e6127-e54b-453f-95b5-a452861e352d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.415815] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 933.415815] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523c6c1e-70cc-c6e1-8479-75c53c288d17" [ 933.415815] env[61867]: _type = "Task" [ 933.415815] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.427958] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523c6c1e-70cc-c6e1-8479-75c53c288d17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.564030] env[61867]: DEBUG nova.objects.base [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Object Instance<9e1cbfa9-28e9-4bca-adfd-78bf25428106> lazy-loaded attributes: resources,numa_topology {{(pid=61867) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 933.568362] env[61867]: DEBUG nova.network.neutron [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 933.572062] env[61867]: DEBUG oslo_concurrency.lockutils [req-15d63e63-71e4-421e-8d12-de5c4e4ade0a req-ce45eb6c-2495-4c6f-8a0f-2baa2957c6a4 service nova] Releasing lock "refresh_cache-38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.572062] env[61867]: DEBUG nova.compute.manager [req-15d63e63-71e4-421e-8d12-de5c4e4ade0a req-ce45eb6c-2495-4c6f-8a0f-2baa2957c6a4 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Received event network-changed-db6e6b48-df38-493c-bc63-881519b16b64 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.572062] env[61867]: DEBUG nova.compute.manager [req-15d63e63-71e4-421e-8d12-de5c4e4ade0a req-ce45eb6c-2495-4c6f-8a0f-2baa2957c6a4 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Refreshing instance network info cache due to event network-changed-db6e6b48-df38-493c-bc63-881519b16b64. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 933.572062] env[61867]: DEBUG oslo_concurrency.lockutils [req-15d63e63-71e4-421e-8d12-de5c4e4ade0a req-ce45eb6c-2495-4c6f-8a0f-2baa2957c6a4 service nova] Acquiring lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.572062] env[61867]: DEBUG oslo_concurrency.lockutils [req-15d63e63-71e4-421e-8d12-de5c4e4ade0a req-ce45eb6c-2495-4c6f-8a0f-2baa2957c6a4 service nova] Acquired lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.572460] env[61867]: DEBUG nova.network.neutron [req-15d63e63-71e4-421e-8d12-de5c4e4ade0a req-ce45eb6c-2495-4c6f-8a0f-2baa2957c6a4 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Refreshing network info cache for port db6e6b48-df38-493c-bc63-881519b16b64 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 933.733099] env[61867]: DEBUG nova.network.neutron [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Updating instance_info_cache with network_info: [{"id": "1a175a33-e511-4f6a-a971-708994d51259", "address": "fa:16:3e:2a:0f:ef", "network": {"id": "9c6e0d78-8bbe-46b4-9233-94f991e940e8", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1003528124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05dd2be830f9410fbb90415cc13ff6bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a175a33-e5", "ovs_interfaceid": "1a175a33-e511-4f6a-a971-708994d51259", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.850800] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc604245-c74a-4305-a19a-6eec63b75302 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.864795] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276899, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.869461] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78adc901-aa61-4004-8f4b-d89598c948db {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.906989] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ace3a6-f52c-4023-a74d-48bebc77ae6b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.915068] env[61867]: DEBUG oslo_vmware.api [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276903, 'name': PowerOnVM_Task, 'duration_secs': 0.449268} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.917323] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 933.922988] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04445bdf-16f5-4fd3-81a6-3f6b5375c8f1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.927702] env[61867]: DEBUG nova.compute.manager [None req-5246bc78-a87c-41fc-b772-0e02c3f5efe8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 933.930209] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645b44c7-6e0b-464a-b42a-934ac8aa3095 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.943083] env[61867]: DEBUG nova.compute.provider_tree [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.951232] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523c6c1e-70cc-c6e1-8479-75c53c288d17, 'name': SearchDatastore_Task, 'duration_secs': 0.0108} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.951892] env[61867]: DEBUG oslo_concurrency.lockutils [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.952146] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 933.952386] env[61867]: DEBUG oslo_concurrency.lockutils [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.952533] env[61867]: DEBUG oslo_concurrency.lockutils [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.952714] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 933.952981] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a9427ad-c1d6-462b-90e6-3941bcd39f8f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.963553] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 933.963760] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 933.965989] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91e3374a-eb7d-4b4d-afec-c00554f923d7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.972363] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 933.972363] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5293934e-a9d3-97de-c1c5-ec8768e08bcd" [ 933.972363] env[61867]: _type = "Task" [ 933.972363] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.984172] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5293934e-a9d3-97de-c1c5-ec8768e08bcd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.206505] env[61867]: DEBUG nova.network.neutron [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Updating instance_info_cache with network_info: [{"id": "254a9643-f941-493e-8c87-2932a6cc00ce", "address": "fa:16:3e:85:01:b4", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap254a9643-f9", "ovs_interfaceid": "254a9643-f941-493e-8c87-2932a6cc00ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.240656] env[61867]: DEBUG oslo_concurrency.lockutils [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Releasing lock "refresh_cache-3905193e-04da-439a-bf6c-16f638a692bf" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.240968] env[61867]: DEBUG nova.compute.manager [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Instance network_info: |[{"id": "1a175a33-e511-4f6a-a971-708994d51259", "address": "fa:16:3e:2a:0f:ef", "network": {"id": "9c6e0d78-8bbe-46b4-9233-94f991e940e8", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1003528124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05dd2be830f9410fbb90415cc13ff6bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a175a33-e5", "ovs_interfaceid": "1a175a33-e511-4f6a-a971-708994d51259", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 934.241411] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:0f:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1a175a33-e511-4f6a-a971-708994d51259', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 934.250277] env[61867]: DEBUG oslo.service.loopingcall [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.253781] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 934.254932] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e56df7d0-a782-40a0-be15-d2214381e9bd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.272753] env[61867]: DEBUG nova.network.neutron [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updating instance_info_cache with network_info: [{"id": "f0eb4126-40e7-4fe8-b276-192b91388aba", "address": "fa:16:3e:69:84:03", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0eb4126-40", "ovs_interfaceid": "f0eb4126-40e7-4fe8-b276-192b91388aba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.281502] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 934.281502] env[61867]: value = "task-1276904" [ 934.281502] env[61867]: _type = "Task" [ 934.281502] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.290896] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276904, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.309058] env[61867]: DEBUG nova.compute.manager [req-cd40d130-8dd7-48c5-ba04-d9736abbdf99 req-2cf2b826-0eb2-4896-b3a4-eedfb265dd18 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Received event network-changed-db6e6b48-df38-493c-bc63-881519b16b64 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.309270] env[61867]: DEBUG nova.compute.manager [req-cd40d130-8dd7-48c5-ba04-d9736abbdf99 req-2cf2b826-0eb2-4896-b3a4-eedfb265dd18 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Refreshing instance network info cache due to event network-changed-db6e6b48-df38-493c-bc63-881519b16b64. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 934.309466] env[61867]: DEBUG oslo_concurrency.lockutils [req-cd40d130-8dd7-48c5-ba04-d9736abbdf99 req-2cf2b826-0eb2-4896-b3a4-eedfb265dd18 service nova] Acquiring lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.352937] env[61867]: DEBUG nova.network.neutron [req-15d63e63-71e4-421e-8d12-de5c4e4ade0a req-ce45eb6c-2495-4c6f-8a0f-2baa2957c6a4 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Updated VIF entry in instance network info cache for port db6e6b48-df38-493c-bc63-881519b16b64. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 934.353526] env[61867]: DEBUG nova.network.neutron [req-15d63e63-71e4-421e-8d12-de5c4e4ade0a req-ce45eb6c-2495-4c6f-8a0f-2baa2957c6a4 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Updating instance_info_cache with network_info: [{"id": "db6e6b48-df38-493c-bc63-881519b16b64", "address": "fa:16:3e:da:a7:2a", "network": {"id": "d8a081a0-0d77-4ae7-9d58-19da456f7699", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-201249363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dca1c24993044e499ef3ded17361258", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb6e6b48-df", "ovs_interfaceid": "db6e6b48-df38-493c-bc63-881519b16b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.365012] env[61867]: DEBUG oslo_vmware.api [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276899, 'name': PowerOnVM_Task, 'duration_secs': 1.30675} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.365299] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 934.365512] env[61867]: INFO nova.compute.manager [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Took 8.40 seconds to spawn the instance on the hypervisor. [ 934.365694] env[61867]: DEBUG nova.compute.manager [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 934.366500] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-691fecc4-b257-4dd0-93cb-859db2d20053 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.455028] env[61867]: DEBUG nova.scheduler.client.report [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 934.484020] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5293934e-a9d3-97de-c1c5-ec8768e08bcd, 'name': SearchDatastore_Task, 'duration_secs': 0.014371} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.485019] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8bcb382-3d16-4243-9fe1-e79660696b3b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.490649] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 934.490649] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522b2881-bd1e-fb26-874f-83a9c28dba03" [ 934.490649] env[61867]: _type = "Task" [ 934.490649] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.500436] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522b2881-bd1e-fb26-874f-83a9c28dba03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.709393] env[61867]: DEBUG oslo_concurrency.lockutils [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "refresh_cache-705a32b6-67f9-42cc-b4d0-f6d1783c68b5" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.709911] env[61867]: DEBUG nova.objects.instance [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lazy-loading 'migration_context' on Instance uuid 705a32b6-67f9-42cc-b4d0-f6d1783c68b5 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.775622] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.794623] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276904, 'name': CreateVM_Task} progress is 25%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.853991] env[61867]: DEBUG nova.compute.manager [req-98bb36dc-5ff8-4a91-8887-22a5fe0d6000 req-6742ae5a-faea-4da4-bd00-6d9085017c95 service nova] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Received event network-changed-1a175a33-e511-4f6a-a971-708994d51259 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.854298] env[61867]: DEBUG nova.compute.manager [req-98bb36dc-5ff8-4a91-8887-22a5fe0d6000 req-6742ae5a-faea-4da4-bd00-6d9085017c95 service nova] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Refreshing instance network info cache due to event network-changed-1a175a33-e511-4f6a-a971-708994d51259. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 934.854425] env[61867]: DEBUG oslo_concurrency.lockutils [req-98bb36dc-5ff8-4a91-8887-22a5fe0d6000 req-6742ae5a-faea-4da4-bd00-6d9085017c95 service nova] Acquiring lock "refresh_cache-3905193e-04da-439a-bf6c-16f638a692bf" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.854577] env[61867]: DEBUG oslo_concurrency.lockutils [req-98bb36dc-5ff8-4a91-8887-22a5fe0d6000 req-6742ae5a-faea-4da4-bd00-6d9085017c95 service nova] Acquired lock "refresh_cache-3905193e-04da-439a-bf6c-16f638a692bf" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.855109] env[61867]: DEBUG nova.network.neutron [req-98bb36dc-5ff8-4a91-8887-22a5fe0d6000 req-6742ae5a-faea-4da4-bd00-6d9085017c95 service nova] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Refreshing network info cache for port 1a175a33-e511-4f6a-a971-708994d51259 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 934.862467] env[61867]: DEBUG oslo_concurrency.lockutils [req-15d63e63-71e4-421e-8d12-de5c4e4ade0a req-ce45eb6c-2495-4c6f-8a0f-2baa2957c6a4 service nova] Releasing lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.863206] env[61867]: DEBUG oslo_concurrency.lockutils [req-cd40d130-8dd7-48c5-ba04-d9736abbdf99 req-2cf2b826-0eb2-4896-b3a4-eedfb265dd18 service nova] Acquired lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.863392] env[61867]: DEBUG nova.network.neutron [req-cd40d130-8dd7-48c5-ba04-d9736abbdf99 req-2cf2b826-0eb2-4896-b3a4-eedfb265dd18 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Refreshing network info cache for port db6e6b48-df38-493c-bc63-881519b16b64 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 934.887442] env[61867]: INFO nova.compute.manager [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Took 16.49 seconds to build instance. [ 934.959775] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.414s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.963448] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.862s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.963689] env[61867]: DEBUG nova.objects.instance [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lazy-loading 'resources' on Instance uuid adc2732d-2a10-40ce-bb90-ed0762a36614 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.002225] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522b2881-bd1e-fb26-874f-83a9c28dba03, 'name': SearchDatastore_Task, 'duration_secs': 0.014287} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.002995] env[61867]: DEBUG oslo_concurrency.lockutils [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.003292] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f/38edb89a-28e4-4dd2-a8aa-35cb95a72e0f.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 935.003790] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d66beef1-0f38-4c98-9f16-cbf11b23b98a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.012053] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 935.012053] env[61867]: value = "task-1276905" [ 935.012053] env[61867]: _type = "Task" [ 935.012053] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.022411] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276905, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.214270] env[61867]: DEBUG nova.objects.base [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Object Instance<705a32b6-67f9-42cc-b4d0-f6d1783c68b5> lazy-loaded attributes: info_cache,migration_context {{(pid=61867) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 935.215773] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb49f76b-a233-422a-a8d1-4f66ce60a644 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.247537] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecbb7f49-6222-497e-a90e-dbe189cd0149 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.254823] env[61867]: DEBUG oslo_vmware.api [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 935.254823] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e8016e-811d-1b6c-89b8-1712593b935b" [ 935.254823] env[61867]: _type = "Task" [ 935.254823] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.264956] env[61867]: DEBUG oslo_vmware.api [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e8016e-811d-1b6c-89b8-1712593b935b, 'name': SearchDatastore_Task, 'duration_secs': 0.0075} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.265286] env[61867]: DEBUG oslo_concurrency.lockutils [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.293365] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276904, 'name': CreateVM_Task, 'duration_secs': 0.674008} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.293627] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 935.294319] env[61867]: DEBUG oslo_concurrency.lockutils [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.294572] env[61867]: DEBUG oslo_concurrency.lockutils [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.295565] env[61867]: DEBUG oslo_concurrency.lockutils [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 935.295565] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12953962-7dfc-421f-aad8-aa604fc47fbf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.304205] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 935.304205] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5231190c-b2df-503a-9f93-ccab613caec2" [ 935.304205] env[61867]: _type = "Task" [ 935.304205] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.316960] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5231190c-b2df-503a-9f93-ccab613caec2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.393195] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cbce68ee-48d7-4fcd-8ce3-051881912f95 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "8a7f4314-0fd4-49f9-8eb6-12baa0977a53" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.011s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.472471] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4eddb87f-8c55-4eae-b186-aea55b1e2e73 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 26.154s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.474403] env[61867]: DEBUG oslo_concurrency.lockutils [None req-633e67b6-a50b-4c01-af8d-8044424e040e tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 5.269s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.474575] env[61867]: DEBUG oslo_concurrency.lockutils [None req-633e67b6-a50b-4c01-af8d-8044424e040e tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.474821] env[61867]: DEBUG oslo_concurrency.lockutils [None req-633e67b6-a50b-4c01-af8d-8044424e040e tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.474995] env[61867]: DEBUG oslo_concurrency.lockutils [None req-633e67b6-a50b-4c01-af8d-8044424e040e tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.481479] env[61867]: INFO nova.compute.manager [None req-633e67b6-a50b-4c01-af8d-8044424e040e tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Terminating instance [ 935.484039] env[61867]: DEBUG nova.compute.manager [None req-633e67b6-a50b-4c01-af8d-8044424e040e tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 935.484326] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-633e67b6-a50b-4c01-af8d-8044424e040e tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 935.484634] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-917f8023-7608-4eb2-b9e3-6a993c780dd5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.506349] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9271abb-d329-459c-86e7-cef0b655ee30 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.531259] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276905, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.547892] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-633e67b6-a50b-4c01-af8d-8044424e040e tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9e1cbfa9-28e9-4bca-adfd-78bf25428106 could not be found. [ 935.548180] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-633e67b6-a50b-4c01-af8d-8044424e040e tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.548438] env[61867]: INFO nova.compute.manager [None req-633e67b6-a50b-4c01-af8d-8044424e040e tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Took 0.06 seconds to destroy the instance on the hypervisor. [ 935.548582] env[61867]: DEBUG oslo.service.loopingcall [None req-633e67b6-a50b-4c01-af8d-8044424e040e tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.551366] env[61867]: DEBUG nova.compute.manager [-] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 935.551502] env[61867]: DEBUG nova.network.neutron [-] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 935.665938] env[61867]: DEBUG nova.network.neutron [req-98bb36dc-5ff8-4a91-8887-22a5fe0d6000 req-6742ae5a-faea-4da4-bd00-6d9085017c95 service nova] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Updated VIF entry in instance network info cache for port 1a175a33-e511-4f6a-a971-708994d51259. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 935.666344] env[61867]: DEBUG nova.network.neutron [req-98bb36dc-5ff8-4a91-8887-22a5fe0d6000 req-6742ae5a-faea-4da4-bd00-6d9085017c95 service nova] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Updating instance_info_cache with network_info: [{"id": "1a175a33-e511-4f6a-a971-708994d51259", "address": "fa:16:3e:2a:0f:ef", "network": {"id": "9c6e0d78-8bbe-46b4-9233-94f991e940e8", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1003528124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05dd2be830f9410fbb90415cc13ff6bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a175a33-e5", "ovs_interfaceid": "1a175a33-e511-4f6a-a971-708994d51259", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.722576] env[61867]: INFO nova.compute.manager [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Unrescuing [ 935.723016] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.723310] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquired lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.723500] env[61867]: DEBUG nova.network.neutron [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 935.756018] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c7f667-3928-4cb9-a0f0-a0075157b7b9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.765044] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89f3e17-2482-4f8d-aca7-91ca9b8d4f4c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.798137] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ba4bab5-1a35-4442-a00d-ca5ade1fe329 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.811818] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed8fe8e-1c80-42a7-b3e1-d5f5f015ad34 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.832021] env[61867]: DEBUG nova.compute.provider_tree [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 935.836493] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5231190c-b2df-503a-9f93-ccab613caec2, 'name': SearchDatastore_Task, 'duration_secs': 0.057458} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.837105] env[61867]: DEBUG oslo_concurrency.lockutils [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.837471] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 935.837796] env[61867]: DEBUG oslo_concurrency.lockutils [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.837997] env[61867]: DEBUG oslo_concurrency.lockutils [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.842020] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 935.842020] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62e36c57-6fcb-4435-a228-0326df6d7e47 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.855755] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 935.855755] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 935.855755] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bd95540-03d9-4d64-aa1b-668b799140e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.865607] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 935.865607] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52aa6d57-22b8-a53a-5abf-87426d770ef3" [ 935.865607] env[61867]: _type = "Task" [ 935.865607] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.876340] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52aa6d57-22b8-a53a-5abf-87426d770ef3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.885232] env[61867]: DEBUG nova.network.neutron [req-cd40d130-8dd7-48c5-ba04-d9736abbdf99 req-2cf2b826-0eb2-4896-b3a4-eedfb265dd18 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Updated VIF entry in instance network info cache for port db6e6b48-df38-493c-bc63-881519b16b64. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 935.885635] env[61867]: DEBUG nova.network.neutron [req-cd40d130-8dd7-48c5-ba04-d9736abbdf99 req-2cf2b826-0eb2-4896-b3a4-eedfb265dd18 service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Updating instance_info_cache with network_info: [{"id": "db6e6b48-df38-493c-bc63-881519b16b64", "address": "fa:16:3e:da:a7:2a", "network": {"id": "d8a081a0-0d77-4ae7-9d58-19da456f7699", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-201249363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dca1c24993044e499ef3ded17361258", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb6e6b48-df", "ovs_interfaceid": "db6e6b48-df38-493c-bc63-881519b16b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.030654] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276905, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.601734} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.030932] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f/38edb89a-28e4-4dd2-a8aa-35cb95a72e0f.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 936.031185] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 936.031458] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6467317a-7da3-4cc2-9db6-ad0932b21fc8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.040264] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 936.040264] env[61867]: value = "task-1276906" [ 936.040264] env[61867]: _type = "Task" [ 936.040264] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.052368] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276906, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.169424] env[61867]: DEBUG oslo_concurrency.lockutils [req-98bb36dc-5ff8-4a91-8887-22a5fe0d6000 req-6742ae5a-faea-4da4-bd00-6d9085017c95 service nova] Releasing lock "refresh_cache-3905193e-04da-439a-bf6c-16f638a692bf" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.305503] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-185140ea-b775-4975-829e-b03849c6a21a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.326120] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updating instance 'abb41c0c-6d0d-4147-a4af-554ab7d9e921' progress to 0 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 936.338085] env[61867]: DEBUG nova.compute.manager [req-4cb00feb-b16c-4b48-9847-76a108c03558 req-8093d6c0-0727-4dec-9780-37b2c09cf5a2 service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Received event network-changed-690a5747-8fcf-445f-9c80-198bfc4fef9a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.338451] env[61867]: DEBUG nova.compute.manager [req-4cb00feb-b16c-4b48-9847-76a108c03558 req-8093d6c0-0727-4dec-9780-37b2c09cf5a2 service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Refreshing instance network info cache due to event network-changed-690a5747-8fcf-445f-9c80-198bfc4fef9a. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 936.338506] env[61867]: DEBUG oslo_concurrency.lockutils [req-4cb00feb-b16c-4b48-9847-76a108c03558 req-8093d6c0-0727-4dec-9780-37b2c09cf5a2 service nova] Acquiring lock "refresh_cache-6d20f463-9198-4590-8ec2-db471c64ba7c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.338693] env[61867]: DEBUG oslo_concurrency.lockutils [req-4cb00feb-b16c-4b48-9847-76a108c03558 req-8093d6c0-0727-4dec-9780-37b2c09cf5a2 service nova] Acquired lock "refresh_cache-6d20f463-9198-4590-8ec2-db471c64ba7c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.338846] env[61867]: DEBUG nova.network.neutron [req-4cb00feb-b16c-4b48-9847-76a108c03558 req-8093d6c0-0727-4dec-9780-37b2c09cf5a2 service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Refreshing network info cache for port 690a5747-8fcf-445f-9c80-198bfc4fef9a {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 936.340277] env[61867]: DEBUG nova.scheduler.client.report [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 936.376933] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52aa6d57-22b8-a53a-5abf-87426d770ef3, 'name': SearchDatastore_Task, 'duration_secs': 0.011395} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.377965] env[61867]: DEBUG nova.network.neutron [-] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.379215] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f64de16-74f8-4d4a-af14-ab3c4d3cea4b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.386553] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 936.386553] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52977b9e-13b5-91f3-c6ac-7eddcda4220e" [ 936.386553] env[61867]: _type = "Task" [ 936.386553] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.390933] env[61867]: DEBUG oslo_concurrency.lockutils [req-cd40d130-8dd7-48c5-ba04-d9736abbdf99 req-2cf2b826-0eb2-4896-b3a4-eedfb265dd18 service nova] Releasing lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.391199] env[61867]: DEBUG nova.compute.manager [req-cd40d130-8dd7-48c5-ba04-d9736abbdf99 req-2cf2b826-0eb2-4896-b3a4-eedfb265dd18 service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Received event network-changed-690a5747-8fcf-445f-9c80-198bfc4fef9a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.391374] env[61867]: DEBUG nova.compute.manager [req-cd40d130-8dd7-48c5-ba04-d9736abbdf99 req-2cf2b826-0eb2-4896-b3a4-eedfb265dd18 service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Refreshing instance network info cache due to event network-changed-690a5747-8fcf-445f-9c80-198bfc4fef9a. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 936.391556] env[61867]: DEBUG oslo_concurrency.lockutils [req-cd40d130-8dd7-48c5-ba04-d9736abbdf99 req-2cf2b826-0eb2-4896-b3a4-eedfb265dd18 service nova] Acquiring lock "refresh_cache-6d20f463-9198-4590-8ec2-db471c64ba7c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.397377] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52977b9e-13b5-91f3-c6ac-7eddcda4220e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.550985] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276906, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08043} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.551292] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 936.552277] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed05fb93-6e0f-46f1-b53a-02655232b99b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.579971] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f/38edb89a-28e4-4dd2-a8aa-35cb95a72e0f.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.582618] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57456767-90d7-4e96-a631-63122d7f29be {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.602829] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 936.602829] env[61867]: value = "task-1276907" [ 936.602829] env[61867]: _type = "Task" [ 936.602829] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.614239] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276907, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.733597] env[61867]: DEBUG oslo_concurrency.lockutils [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "6d20f463-9198-4590-8ec2-db471c64ba7c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.733887] env[61867]: DEBUG oslo_concurrency.lockutils [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "6d20f463-9198-4590-8ec2-db471c64ba7c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.734181] env[61867]: DEBUG oslo_concurrency.lockutils [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "6d20f463-9198-4590-8ec2-db471c64ba7c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.734389] env[61867]: DEBUG oslo_concurrency.lockutils [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "6d20f463-9198-4590-8ec2-db471c64ba7c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.734568] env[61867]: DEBUG oslo_concurrency.lockutils [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "6d20f463-9198-4590-8ec2-db471c64ba7c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.737668] env[61867]: INFO nova.compute.manager [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Terminating instance [ 936.740026] env[61867]: DEBUG nova.compute.manager [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 936.740230] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 936.741117] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196fd165-cded-4793-8244-57f9f7eef25b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.749614] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 936.749860] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7fdfb397-4729-42f8-a7ef-be7a50b21c31 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.757040] env[61867]: DEBUG oslo_vmware.api [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 936.757040] env[61867]: value = "task-1276908" [ 936.757040] env[61867]: _type = "Task" [ 936.757040] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.765662] env[61867]: DEBUG oslo_vmware.api [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276908, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.831916] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 936.832375] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96a00548-03d6-4179-84a9-f9c147e11ccc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.842101] env[61867]: DEBUG oslo_vmware.api [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 936.842101] env[61867]: value = "task-1276909" [ 936.842101] env[61867]: _type = "Task" [ 936.842101] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.850098] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.887s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.852292] env[61867]: DEBUG oslo_concurrency.lockutils [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.541s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.853866] env[61867]: INFO nova.compute.claims [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 936.866538] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] VM already powered off {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 936.866849] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updating instance 'abb41c0c-6d0d-4147-a4af-554ab7d9e921' progress to 17 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 936.880808] env[61867]: INFO nova.compute.manager [-] [instance: 9e1cbfa9-28e9-4bca-adfd-78bf25428106] Took 1.33 seconds to deallocate network for instance. [ 936.889192] env[61867]: INFO nova.scheduler.client.report [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleted allocations for instance adc2732d-2a10-40ce-bb90-ed0762a36614 [ 936.904849] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52977b9e-13b5-91f3-c6ac-7eddcda4220e, 'name': SearchDatastore_Task, 'duration_secs': 0.054046} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.905239] env[61867]: DEBUG oslo_concurrency.lockutils [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.905395] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 3905193e-04da-439a-bf6c-16f638a692bf/3905193e-04da-439a-bf6c-16f638a692bf.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 936.905666] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5fadb666-4d9a-4199-b223-11ddcb8e1c1a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.914303] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 936.914303] env[61867]: value = "task-1276910" [ 936.914303] env[61867]: _type = "Task" [ 936.914303] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.924373] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276910, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.087083] env[61867]: DEBUG nova.network.neutron [req-4cb00feb-b16c-4b48-9847-76a108c03558 req-8093d6c0-0727-4dec-9780-37b2c09cf5a2 service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Updated VIF entry in instance network info cache for port 690a5747-8fcf-445f-9c80-198bfc4fef9a. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 937.087526] env[61867]: DEBUG nova.network.neutron [req-4cb00feb-b16c-4b48-9847-76a108c03558 req-8093d6c0-0727-4dec-9780-37b2c09cf5a2 service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Updating instance_info_cache with network_info: [{"id": "690a5747-8fcf-445f-9c80-198bfc4fef9a", "address": "fa:16:3e:99:85:22", "network": {"id": "d8a081a0-0d77-4ae7-9d58-19da456f7699", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-201249363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dca1c24993044e499ef3ded17361258", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap690a5747-8f", "ovs_interfaceid": "690a5747-8fcf-445f-9c80-198bfc4fef9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.105648] env[61867]: DEBUG nova.network.neutron [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Updating instance_info_cache with network_info: [{"id": "4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1", "address": "fa:16:3e:ce:d6:b5", "network": {"id": "ab6ec5b9-8193-43c9-a3ce-d63f31e5a001", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1573384141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35aa85105b394369b17ef4ce51a49066", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f63bd0e-c1", "ovs_interfaceid": "4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.118880] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276907, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.275211] env[61867]: DEBUG oslo_vmware.api [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276908, 'name': PowerOffVM_Task, 'duration_secs': 0.215797} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.275211] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 937.277546] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 937.277950] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f7f23b5-a956-4807-b740-95782fefe4bf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.377558] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 937.378176] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 937.378425] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 937.379293] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 937.379293] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 937.379293] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 937.379603] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 937.379832] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 937.381473] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 937.381805] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 937.382079] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 937.390472] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5385874-6702-4db4-be8c-47a89c4a5012 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.402361] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 937.402638] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 937.403531] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Deleting the datastore file [datastore2] 6d20f463-9198-4590-8ec2-db471c64ba7c {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 937.410315] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e216861-6ec1-4681-baf3-287eb7e8a776 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.413461] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3732613c-ec66-4c74-beb4-eb93fdc735f8 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "adc2732d-2a10-40ce-bb90-ed0762a36614" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.401s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.422451] env[61867]: DEBUG oslo_vmware.api [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 937.422451] env[61867]: value = "task-1276912" [ 937.422451] env[61867]: _type = "Task" [ 937.422451] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.425088] env[61867]: DEBUG oslo_vmware.api [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 937.425088] env[61867]: value = "task-1276913" [ 937.425088] env[61867]: _type = "Task" [ 937.425088] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.434399] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276910, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.454289] env[61867]: DEBUG oslo_vmware.api [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276912, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.457132] env[61867]: DEBUG oslo_vmware.api [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276913, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.590955] env[61867]: DEBUG oslo_concurrency.lockutils [req-4cb00feb-b16c-4b48-9847-76a108c03558 req-8093d6c0-0727-4dec-9780-37b2c09cf5a2 service nova] Releasing lock "refresh_cache-6d20f463-9198-4590-8ec2-db471c64ba7c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.591848] env[61867]: DEBUG oslo_concurrency.lockutils [req-cd40d130-8dd7-48c5-ba04-d9736abbdf99 req-2cf2b826-0eb2-4896-b3a4-eedfb265dd18 service nova] Acquired lock "refresh_cache-6d20f463-9198-4590-8ec2-db471c64ba7c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.592151] env[61867]: DEBUG nova.network.neutron [req-cd40d130-8dd7-48c5-ba04-d9736abbdf99 req-2cf2b826-0eb2-4896-b3a4-eedfb265dd18 service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Refreshing network info cache for port 690a5747-8fcf-445f-9c80-198bfc4fef9a {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 937.608259] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Releasing lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.608944] env[61867]: DEBUG nova.objects.instance [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lazy-loading 'flavor' on Instance uuid 7479bf91-5aef-4e75-a127-7e82ae15a003 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.622301] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276907, 'name': ReconfigVM_Task, 'duration_secs': 0.672698} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.623240] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f/38edb89a-28e4-4dd2-a8aa-35cb95a72e0f.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.623883] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-10a2f4b6-dee7-4c2b-a1a8-1ab7cdfe934e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.631861] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 937.631861] env[61867]: value = "task-1276914" [ 937.631861] env[61867]: _type = "Task" [ 937.631861] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.641032] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276914, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.931988] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276910, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.638673} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.942022] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 3905193e-04da-439a-bf6c-16f638a692bf/3905193e-04da-439a-bf6c-16f638a692bf.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 937.942022] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 937.942022] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-47c0ed43-a3c8-426a-850f-4adc587f5fe9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.954027] env[61867]: DEBUG oslo_concurrency.lockutils [None req-633e67b6-a50b-4c01-af8d-8044424e040e tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "9e1cbfa9-28e9-4bca-adfd-78bf25428106" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.480s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.955163] env[61867]: DEBUG oslo_vmware.api [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276912, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.303454} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.964381] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 937.964763] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 937.965094] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 937.965373] env[61867]: INFO nova.compute.manager [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Took 1.23 seconds to destroy the instance on the hypervisor. [ 937.965700] env[61867]: DEBUG oslo.service.loopingcall [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.966038] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 937.966038] env[61867]: value = "task-1276915" [ 937.966038] env[61867]: _type = "Task" [ 937.966038] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.966288] env[61867]: DEBUG oslo_vmware.api [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276913, 'name': ReconfigVM_Task, 'duration_secs': 0.203823} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.967870] env[61867]: DEBUG nova.compute.manager [-] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 937.967979] env[61867]: DEBUG nova.network.neutron [-] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 937.969808] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updating instance 'abb41c0c-6d0d-4147-a4af-554ab7d9e921' progress to 33 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 937.985149] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276915, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.118406] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da835245-bcd7-458d-8148-aeb7ed4692ec {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.151404] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 938.159306] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cdab32db-5e77-4133-bf54-4ee8036689ac {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.170883] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276914, 'name': Rename_Task, 'duration_secs': 0.320544} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.172426] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 938.172825] env[61867]: DEBUG oslo_vmware.api [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 938.172825] env[61867]: value = "task-1276916" [ 938.172825] env[61867]: _type = "Task" [ 938.172825] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.173098] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1ddfce0d-e0d2-4aef-8b32-47b15ab307d7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.180424] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa37192-b0c6-4b2e-854a-d74beb6628d9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.185483] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 938.185483] env[61867]: value = "task-1276917" [ 938.185483] env[61867]: _type = "Task" [ 938.185483] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.191751] env[61867]: DEBUG oslo_vmware.api [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276916, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.196046] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae023960-6106-41a8-97a6-f389dca12822 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.201806] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276917, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.232172] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90225db8-4f0f-4baa-a158-f947d7ffa95e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.240733] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f86682b-4e15-4b36-ae19-ac26b087f368 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.255575] env[61867]: DEBUG nova.compute.provider_tree [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.402076] env[61867]: DEBUG nova.network.neutron [req-cd40d130-8dd7-48c5-ba04-d9736abbdf99 req-2cf2b826-0eb2-4896-b3a4-eedfb265dd18 service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Updated VIF entry in instance network info cache for port 690a5747-8fcf-445f-9c80-198bfc4fef9a. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 938.402466] env[61867]: DEBUG nova.network.neutron [req-cd40d130-8dd7-48c5-ba04-d9736abbdf99 req-2cf2b826-0eb2-4896-b3a4-eedfb265dd18 service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Updating instance_info_cache with network_info: [{"id": "690a5747-8fcf-445f-9c80-198bfc4fef9a", "address": "fa:16:3e:99:85:22", "network": {"id": "d8a081a0-0d77-4ae7-9d58-19da456f7699", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-201249363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dca1c24993044e499ef3ded17361258", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap690a5747-8f", "ovs_interfaceid": "690a5747-8fcf-445f-9c80-198bfc4fef9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.476768] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 938.477067] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 938.477269] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 938.477743] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 938.477901] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 938.478120] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 938.478381] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 938.478614] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 938.478826] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 938.479055] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 938.479282] env[61867]: DEBUG nova.virt.hardware [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 938.485914] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Reconfiguring VM instance instance-00000042 to detach disk 2000 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 938.486220] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6921bcad-4b64-4f68-959b-925bf8704521 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.508381] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276915, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07664} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.509766] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 938.510174] env[61867]: DEBUG oslo_vmware.api [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 938.510174] env[61867]: value = "task-1276918" [ 938.510174] env[61867]: _type = "Task" [ 938.510174] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.510878] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27602816-e580-485b-a1dd-aa91edfe1d34 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.524826] env[61867]: DEBUG oslo_vmware.api [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276918, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.543787] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 3905193e-04da-439a-bf6c-16f638a692bf/3905193e-04da-439a-bf6c-16f638a692bf.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 938.544911] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66261f51-95b8-46f6-9a57-addfc351a33b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.565686] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 938.565686] env[61867]: value = "task-1276919" [ 938.565686] env[61867]: _type = "Task" [ 938.565686] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.575179] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276919, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.687215] env[61867]: DEBUG oslo_vmware.api [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276916, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.696986] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276917, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.760206] env[61867]: DEBUG nova.scheduler.client.report [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 938.867046] env[61867]: DEBUG nova.compute.manager [req-0d0326bf-2e39-4c88-b8c9-04c15b523b66 req-fa2613c1-a0eb-4449-be44-b7123c54a1ab service nova] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Received event network-changed-211a1ec3-d618-4c1d-ac61-7e7b5681b9ed {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.867274] env[61867]: DEBUG nova.compute.manager [req-0d0326bf-2e39-4c88-b8c9-04c15b523b66 req-fa2613c1-a0eb-4449-be44-b7123c54a1ab service nova] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Refreshing instance network info cache due to event network-changed-211a1ec3-d618-4c1d-ac61-7e7b5681b9ed. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 938.867492] env[61867]: DEBUG oslo_concurrency.lockutils [req-0d0326bf-2e39-4c88-b8c9-04c15b523b66 req-fa2613c1-a0eb-4449-be44-b7123c54a1ab service nova] Acquiring lock "refresh_cache-8a7f4314-0fd4-49f9-8eb6-12baa0977a53" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.867677] env[61867]: DEBUG oslo_concurrency.lockutils [req-0d0326bf-2e39-4c88-b8c9-04c15b523b66 req-fa2613c1-a0eb-4449-be44-b7123c54a1ab service nova] Acquired lock "refresh_cache-8a7f4314-0fd4-49f9-8eb6-12baa0977a53" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.867899] env[61867]: DEBUG nova.network.neutron [req-0d0326bf-2e39-4c88-b8c9-04c15b523b66 req-fa2613c1-a0eb-4449-be44-b7123c54a1ab service nova] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Refreshing network info cache for port 211a1ec3-d618-4c1d-ac61-7e7b5681b9ed {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 938.905591] env[61867]: DEBUG oslo_concurrency.lockutils [req-cd40d130-8dd7-48c5-ba04-d9736abbdf99 req-2cf2b826-0eb2-4896-b3a4-eedfb265dd18 service nova] Releasing lock "refresh_cache-6d20f463-9198-4590-8ec2-db471c64ba7c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.024945] env[61867]: DEBUG oslo_vmware.api [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276918, 'name': ReconfigVM_Task, 'duration_secs': 0.398124} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.025302] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Reconfigured VM instance instance-00000042 to detach disk 2000 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 939.026077] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e3d519f-78a5-42cc-a7e2-3c313783adab {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.041974] env[61867]: DEBUG nova.network.neutron [-] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.050505] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] abb41c0c-6d0d-4147-a4af-554ab7d9e921/abb41c0c-6d0d-4147-a4af-554ab7d9e921.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.051120] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57537f24-979a-49f5-b4d6-94102993dad7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.072414] env[61867]: DEBUG oslo_vmware.api [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 939.072414] env[61867]: value = "task-1276920" [ 939.072414] env[61867]: _type = "Task" [ 939.072414] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.080354] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276919, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.086240] env[61867]: DEBUG oslo_vmware.api [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276920, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.190026] env[61867]: DEBUG oslo_vmware.api [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276916, 'name': PowerOffVM_Task, 'duration_secs': 0.5276} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.193865] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 939.200877] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Reconfiguring VM instance instance-00000050 to detach disk 2002 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 939.201405] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6572212-4a56-4664-bab4-9f5741b8bfcd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.230951] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276917, 'name': PowerOnVM_Task} progress is 81%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.236025] env[61867]: DEBUG oslo_vmware.api [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 939.236025] env[61867]: value = "task-1276921" [ 939.236025] env[61867]: _type = "Task" [ 939.236025] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.242418] env[61867]: DEBUG oslo_vmware.api [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276921, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.262121] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "43dc6520-5256-4b5f-a273-6c9e0e407c72" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.262121] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "43dc6520-5256-4b5f-a273-6c9e0e407c72" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.265336] env[61867]: DEBUG oslo_concurrency.lockutils [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.266059] env[61867]: DEBUG nova.compute.manager [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 939.268935] env[61867]: DEBUG oslo_concurrency.lockutils [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "5cf15593-54f6-405c-8435-2e3b378983b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.269349] env[61867]: DEBUG oslo_concurrency.lockutils [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "5cf15593-54f6-405c-8435-2e3b378983b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.270724] env[61867]: DEBUG oslo_concurrency.lockutils [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 4.005s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.554596] env[61867]: INFO nova.compute.manager [-] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Took 1.59 seconds to deallocate network for instance. [ 939.585732] env[61867]: DEBUG oslo_vmware.api [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276920, 'name': ReconfigVM_Task, 'duration_secs': 0.332439} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.590096] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Reconfigured VM instance instance-00000042 to attach disk [datastore1] abb41c0c-6d0d-4147-a4af-554ab7d9e921/abb41c0c-6d0d-4147-a4af-554ab7d9e921.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 939.590405] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updating instance 'abb41c0c-6d0d-4147-a4af-554ab7d9e921' progress to 50 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 939.594288] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276919, 'name': ReconfigVM_Task, 'duration_secs': 0.701453} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.598104] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 3905193e-04da-439a-bf6c-16f638a692bf/3905193e-04da-439a-bf6c-16f638a692bf.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 939.598104] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-29ad0918-b324-40ca-9b03-83b19c24383c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.607882] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 939.607882] env[61867]: value = "task-1276922" [ 939.607882] env[61867]: _type = "Task" [ 939.607882] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.618814] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276922, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.700427] env[61867]: DEBUG oslo_vmware.api [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276917, 'name': PowerOnVM_Task, 'duration_secs': 1.446591} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.700692] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 939.700898] env[61867]: INFO nova.compute.manager [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Took 16.13 seconds to spawn the instance on the hypervisor. [ 939.701091] env[61867]: DEBUG nova.compute.manager [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 939.701929] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a8bf6c2-c2a8-464b-834b-efda70250a3f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.743700] env[61867]: DEBUG oslo_vmware.api [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276921, 'name': ReconfigVM_Task, 'duration_secs': 0.422209} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.743945] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Reconfigured VM instance instance-00000050 to detach disk 2002 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 939.744030] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 939.744904] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-071bab13-a712-42f2-a085-7dfb6f5fd6ea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.752548] env[61867]: DEBUG oslo_vmware.api [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 939.752548] env[61867]: value = "task-1276923" [ 939.752548] env[61867]: _type = "Task" [ 939.752548] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.762634] env[61867]: DEBUG oslo_vmware.api [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276923, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.765129] env[61867]: DEBUG nova.compute.manager [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 939.773618] env[61867]: DEBUG nova.compute.utils [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 939.773618] env[61867]: DEBUG nova.compute.manager [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 939.774675] env[61867]: DEBUG nova.network.neutron [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 939.778705] env[61867]: DEBUG nova.compute.manager [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 939.827359] env[61867]: DEBUG nova.policy [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ea634246436422a9407f82692b69ede', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '86ced20eaf4740e298dc6f8ca5550c09', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 939.843578] env[61867]: DEBUG nova.network.neutron [req-0d0326bf-2e39-4c88-b8c9-04c15b523b66 req-fa2613c1-a0eb-4449-be44-b7123c54a1ab service nova] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Updated VIF entry in instance network info cache for port 211a1ec3-d618-4c1d-ac61-7e7b5681b9ed. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 939.845174] env[61867]: DEBUG nova.network.neutron [req-0d0326bf-2e39-4c88-b8c9-04c15b523b66 req-fa2613c1-a0eb-4449-be44-b7123c54a1ab service nova] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Updating instance_info_cache with network_info: [{"id": "211a1ec3-d618-4c1d-ac61-7e7b5681b9ed", "address": "fa:16:3e:b3:dd:4d", "network": {"id": "fc403caf-421f-4783-9807-5e61c4379a6d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-846804042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24180ce8aca142fb897e29d4853c20c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap211a1ec3-d6", "ovs_interfaceid": "211a1ec3-d618-4c1d-ac61-7e7b5681b9ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.067904] env[61867]: DEBUG oslo_concurrency.lockutils [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.078090] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5488599-43db-4ad0-aa9a-589c61e7e32a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.094304] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d30df0b5-a8ff-48ad-98ca-3253aafb1559 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.106550] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c17747ef-fe45-4411-aa29-7504ef55f446 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.167494] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8103a5d-db0d-4cb4-94a7-324ad5853fd8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.175029] env[61867]: DEBUG nova.network.neutron [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Successfully created port: c05702c4-5138-450d-82b2-790b86d8a59b {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 940.177442] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-244014d1-b496-4f1a-a1bf-fc155c3f99e6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.180100] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276922, 'name': Rename_Task, 'duration_secs': 0.299464} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.180794] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 940.181654] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bdb6e679-6e7b-4e10-9759-846569992e86 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.200434] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updating instance 'abb41c0c-6d0d-4147-a4af-554ab7d9e921' progress to 67 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 940.205458] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1888c2a7-125c-47be-87aa-d45bb0569ff9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.210900] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 940.210900] env[61867]: value = "task-1276924" [ 940.210900] env[61867]: _type = "Task" [ 940.210900] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.229019] env[61867]: DEBUG nova.compute.provider_tree [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.231152] env[61867]: INFO nova.compute.manager [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Took 23.81 seconds to build instance. [ 940.235997] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276924, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.263203] env[61867]: DEBUG oslo_vmware.api [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276923, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.279203] env[61867]: DEBUG nova.compute.manager [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 940.288832] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.297930] env[61867]: DEBUG oslo_concurrency.lockutils [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.349097] env[61867]: DEBUG oslo_concurrency.lockutils [req-0d0326bf-2e39-4c88-b8c9-04c15b523b66 req-fa2613c1-a0eb-4449-be44-b7123c54a1ab service nova] Releasing lock "refresh_cache-8a7f4314-0fd4-49f9-8eb6-12baa0977a53" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.349303] env[61867]: DEBUG nova.compute.manager [req-0d0326bf-2e39-4c88-b8c9-04c15b523b66 req-fa2613c1-a0eb-4449-be44-b7123c54a1ab service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Received event network-changed-db6e6b48-df38-493c-bc63-881519b16b64 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.349532] env[61867]: DEBUG nova.compute.manager [req-0d0326bf-2e39-4c88-b8c9-04c15b523b66 req-fa2613c1-a0eb-4449-be44-b7123c54a1ab service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Refreshing instance network info cache due to event network-changed-db6e6b48-df38-493c-bc63-881519b16b64. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 940.349817] env[61867]: DEBUG oslo_concurrency.lockutils [req-0d0326bf-2e39-4c88-b8c9-04c15b523b66 req-fa2613c1-a0eb-4449-be44-b7123c54a1ab service nova] Acquiring lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.349981] env[61867]: DEBUG oslo_concurrency.lockutils [req-0d0326bf-2e39-4c88-b8c9-04c15b523b66 req-fa2613c1-a0eb-4449-be44-b7123c54a1ab service nova] Acquired lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.350170] env[61867]: DEBUG nova.network.neutron [req-0d0326bf-2e39-4c88-b8c9-04c15b523b66 req-fa2613c1-a0eb-4449-be44-b7123c54a1ab service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Refreshing network info cache for port db6e6b48-df38-493c-bc63-881519b16b64 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 940.387905] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "6d20f463-9198-4590-8ec2-db471c64ba7c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.588663] env[61867]: DEBUG oslo_concurrency.lockutils [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquiring lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.724270] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276924, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.732660] env[61867]: DEBUG nova.scheduler.client.report [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 940.737096] env[61867]: DEBUG oslo_concurrency.lockutils [None req-56d9fd88-5ee8-4204-8934-f6505b23efb0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.329s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.737387] env[61867]: DEBUG oslo_concurrency.lockutils [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.149s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.737608] env[61867]: DEBUG oslo_concurrency.lockutils [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquiring lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.737817] env[61867]: DEBUG oslo_concurrency.lockutils [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.737994] env[61867]: DEBUG oslo_concurrency.lockutils [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.740280] env[61867]: INFO nova.compute.manager [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Terminating instance [ 940.742833] env[61867]: DEBUG nova.compute.manager [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 940.743074] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 940.743967] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2fb54c0-ffd0-43e9-b105-aea360d3ffbf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.752797] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 940.753104] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a0eb6e46-0acb-48c2-a4cf-f0c18a5f544d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.755718] env[61867]: DEBUG nova.network.neutron [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Port f0eb4126-40e7-4fe8-b276-192b91388aba binding to destination host cpu-1 is already ACTIVE {{(pid=61867) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 940.766306] env[61867]: DEBUG oslo_vmware.api [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276923, 'name': PowerOnVM_Task, 'duration_secs': 0.650787} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.768210] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 940.768459] env[61867]: DEBUG nova.compute.manager [None req-2a4e3619-acf3-4393-988f-d7e71c6b5d60 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 940.768802] env[61867]: DEBUG oslo_vmware.api [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 940.768802] env[61867]: value = "task-1276925" [ 940.768802] env[61867]: _type = "Task" [ 940.768802] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.769595] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3909389d-6cb9-48f7-abea-0f788a61e40c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.789946] env[61867]: DEBUG oslo_vmware.api [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276925, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.905275] env[61867]: DEBUG nova.compute.manager [req-022bec36-4dfa-4f16-806b-8dd539b6c949 req-edeefa2f-4915-4997-88d5-5094eae4cd3f service nova] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Received event network-vif-deleted-690a5747-8fcf-445f-9c80-198bfc4fef9a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.905506] env[61867]: DEBUG nova.compute.manager [req-022bec36-4dfa-4f16-806b-8dd539b6c949 req-edeefa2f-4915-4997-88d5-5094eae4cd3f service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Received event network-changed-db6e6b48-df38-493c-bc63-881519b16b64 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.905673] env[61867]: DEBUG nova.compute.manager [req-022bec36-4dfa-4f16-806b-8dd539b6c949 req-edeefa2f-4915-4997-88d5-5094eae4cd3f service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Refreshing instance network info cache due to event network-changed-db6e6b48-df38-493c-bc63-881519b16b64. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 940.905864] env[61867]: DEBUG oslo_concurrency.lockutils [req-022bec36-4dfa-4f16-806b-8dd539b6c949 req-edeefa2f-4915-4997-88d5-5094eae4cd3f service nova] Acquiring lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.189616] env[61867]: DEBUG nova.network.neutron [req-0d0326bf-2e39-4c88-b8c9-04c15b523b66 req-fa2613c1-a0eb-4449-be44-b7123c54a1ab service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Updated VIF entry in instance network info cache for port db6e6b48-df38-493c-bc63-881519b16b64. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 941.190211] env[61867]: DEBUG nova.network.neutron [req-0d0326bf-2e39-4c88-b8c9-04c15b523b66 req-fa2613c1-a0eb-4449-be44-b7123c54a1ab service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Updating instance_info_cache with network_info: [{"id": "db6e6b48-df38-493c-bc63-881519b16b64", "address": "fa:16:3e:da:a7:2a", "network": {"id": "d8a081a0-0d77-4ae7-9d58-19da456f7699", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-201249363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dca1c24993044e499ef3ded17361258", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb6e6b48-df", "ovs_interfaceid": "db6e6b48-df38-493c-bc63-881519b16b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.224718] env[61867]: DEBUG oslo_vmware.api [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276924, 'name': PowerOnVM_Task, 'duration_secs': 0.946817} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.225018] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 941.225237] env[61867]: INFO nova.compute.manager [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Took 9.07 seconds to spawn the instance on the hypervisor. [ 941.225427] env[61867]: DEBUG nova.compute.manager [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 941.226299] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af56819-5848-41de-8b47-a18de1280ecb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.283273] env[61867]: DEBUG oslo_vmware.api [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276925, 'name': PowerOffVM_Task, 'duration_secs': 0.288465} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.283504] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 941.283682] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 941.283944] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3db93819-1d76-4797-bedf-0d5c008f78ec {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.292013] env[61867]: DEBUG nova.compute.manager [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 941.322667] env[61867]: DEBUG nova.virt.hardware [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 941.322947] env[61867]: DEBUG nova.virt.hardware [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 941.323144] env[61867]: DEBUG nova.virt.hardware [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 941.323363] env[61867]: DEBUG nova.virt.hardware [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 941.323514] env[61867]: DEBUG nova.virt.hardware [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 941.323667] env[61867]: DEBUG nova.virt.hardware [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 941.323882] env[61867]: DEBUG nova.virt.hardware [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 941.324064] env[61867]: DEBUG nova.virt.hardware [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 941.324243] env[61867]: DEBUG nova.virt.hardware [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 941.324410] env[61867]: DEBUG nova.virt.hardware [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 941.324598] env[61867]: DEBUG nova.virt.hardware [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 941.325908] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2d2c89-3f0d-41ec-a22e-0511077682da {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.334856] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e49c1a8-f973-4546-8e27-a78451410b98 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.442839] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 941.443130] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 941.443343] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Deleting the datastore file [datastore2] 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 941.443609] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4b7f17a-c7a1-4967-94f5-ef2840271e3c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.451376] env[61867]: DEBUG oslo_vmware.api [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 941.451376] env[61867]: value = "task-1276927" [ 941.451376] env[61867]: _type = "Task" [ 941.451376] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.459765] env[61867]: DEBUG oslo_vmware.api [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276927, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.610532] env[61867]: DEBUG nova.compute.manager [req-4e460cc7-4824-4a90-9cb9-e480d47f8b01 req-a8a02361-2a1f-45aa-94b7-f4dc8ac324f8 service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Received event network-vif-plugged-c05702c4-5138-450d-82b2-790b86d8a59b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.610840] env[61867]: DEBUG oslo_concurrency.lockutils [req-4e460cc7-4824-4a90-9cb9-e480d47f8b01 req-a8a02361-2a1f-45aa-94b7-f4dc8ac324f8 service nova] Acquiring lock "4d7bfb31-d565-49e4-8c51-0122acebba2d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.611163] env[61867]: DEBUG oslo_concurrency.lockutils [req-4e460cc7-4824-4a90-9cb9-e480d47f8b01 req-a8a02361-2a1f-45aa-94b7-f4dc8ac324f8 service nova] Lock "4d7bfb31-d565-49e4-8c51-0122acebba2d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.611379] env[61867]: DEBUG oslo_concurrency.lockutils [req-4e460cc7-4824-4a90-9cb9-e480d47f8b01 req-a8a02361-2a1f-45aa-94b7-f4dc8ac324f8 service nova] Lock "4d7bfb31-d565-49e4-8c51-0122acebba2d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.611594] env[61867]: DEBUG nova.compute.manager [req-4e460cc7-4824-4a90-9cb9-e480d47f8b01 req-a8a02361-2a1f-45aa-94b7-f4dc8ac324f8 service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] No waiting events found dispatching network-vif-plugged-c05702c4-5138-450d-82b2-790b86d8a59b {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 941.611900] env[61867]: WARNING nova.compute.manager [req-4e460cc7-4824-4a90-9cb9-e480d47f8b01 req-a8a02361-2a1f-45aa-94b7-f4dc8ac324f8 service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Received unexpected event network-vif-plugged-c05702c4-5138-450d-82b2-790b86d8a59b for instance with vm_state building and task_state spawning. [ 941.693598] env[61867]: DEBUG oslo_concurrency.lockutils [req-0d0326bf-2e39-4c88-b8c9-04c15b523b66 req-fa2613c1-a0eb-4449-be44-b7123c54a1ab service nova] Releasing lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.693598] env[61867]: DEBUG oslo_concurrency.lockutils [req-022bec36-4dfa-4f16-806b-8dd539b6c949 req-edeefa2f-4915-4997-88d5-5094eae4cd3f service nova] Acquired lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.693598] env[61867]: DEBUG nova.network.neutron [req-022bec36-4dfa-4f16-806b-8dd539b6c949 req-edeefa2f-4915-4997-88d5-5094eae4cd3f service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Refreshing network info cache for port db6e6b48-df38-493c-bc63-881519b16b64 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 941.715392] env[61867]: DEBUG nova.network.neutron [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Successfully updated port: c05702c4-5138-450d-82b2-790b86d8a59b {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 941.746589] env[61867]: DEBUG oslo_concurrency.lockutils [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.476s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.752394] env[61867]: INFO nova.compute.manager [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Took 20.61 seconds to build instance. [ 941.754759] env[61867]: DEBUG oslo_concurrency.lockutils [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.688s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.755033] env[61867]: DEBUG nova.objects.instance [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lazy-loading 'resources' on Instance uuid 6d20f463-9198-4590-8ec2-db471c64ba7c {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.773983] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.774268] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.774451] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.964036] env[61867]: DEBUG oslo_vmware.api [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276927, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217399} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.964036] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 941.964036] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 941.964036] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 941.964036] env[61867]: INFO nova.compute.manager [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Took 1.22 seconds to destroy the instance on the hypervisor. [ 941.964355] env[61867]: DEBUG oslo.service.loopingcall [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.964464] env[61867]: DEBUG nova.compute.manager [-] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 941.964561] env[61867]: DEBUG nova.network.neutron [-] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 942.218421] env[61867]: DEBUG oslo_concurrency.lockutils [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.218742] env[61867]: DEBUG oslo_concurrency.lockutils [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.218742] env[61867]: DEBUG nova.network.neutron [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 942.256931] env[61867]: DEBUG oslo_concurrency.lockutils [None req-707e4e39-f048-41ff-9b31-fd6a7d802b42 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "3905193e-04da-439a-bf6c-16f638a692bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.118s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.327869] env[61867]: INFO nova.scheduler.client.report [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleted allocation for migration 328e4a90-96f6-4e27-ba58-da709d0e6338 [ 942.532512] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67176c68-50df-4578-b034-5add1b45bfca {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.540808] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac8fd7c-6865-401b-9c7e-299db58f93c1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.573993] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6cf7d43-3f77-47ec-a194-25bf8a37ea22 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.582939] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b7f035-5497-46f1-90ec-eebd51e55b9c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.601639] env[61867]: DEBUG nova.compute.provider_tree [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.790925] env[61867]: DEBUG nova.network.neutron [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 942.826653] env[61867]: DEBUG nova.network.neutron [req-022bec36-4dfa-4f16-806b-8dd539b6c949 req-edeefa2f-4915-4997-88d5-5094eae4cd3f service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Updated VIF entry in instance network info cache for port db6e6b48-df38-493c-bc63-881519b16b64. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 942.826653] env[61867]: DEBUG nova.network.neutron [req-022bec36-4dfa-4f16-806b-8dd539b6c949 req-edeefa2f-4915-4997-88d5-5094eae4cd3f service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Updating instance_info_cache with network_info: [{"id": "db6e6b48-df38-493c-bc63-881519b16b64", "address": "fa:16:3e:da:a7:2a", "network": {"id": "d8a081a0-0d77-4ae7-9d58-19da456f7699", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-201249363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dca1c24993044e499ef3ded17361258", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb6e6b48-df", "ovs_interfaceid": "db6e6b48-df38-493c-bc63-881519b16b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.838245] env[61867]: DEBUG oslo_concurrency.lockutils [None req-155de411-9087-4992-b62c-ff79abf8e3c8 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 10.536s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.876544] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.876544] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.876544] env[61867]: DEBUG nova.network.neutron [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 943.105570] env[61867]: DEBUG nova.scheduler.client.report [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.205316] env[61867]: DEBUG nova.network.neutron [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Updating instance_info_cache with network_info: [{"id": "c05702c4-5138-450d-82b2-790b86d8a59b", "address": "fa:16:3e:cf:d4:2a", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc05702c4-51", "ovs_interfaceid": "c05702c4-5138-450d-82b2-790b86d8a59b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.330313] env[61867]: DEBUG oslo_concurrency.lockutils [req-022bec36-4dfa-4f16-806b-8dd539b6c949 req-edeefa2f-4915-4997-88d5-5094eae4cd3f service nova] Releasing lock "refresh_cache-816ca796-d8ed-4843-9b1f-f169f48ff047" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.587934] env[61867]: DEBUG nova.network.neutron [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updating instance_info_cache with network_info: [{"id": "f0eb4126-40e7-4fe8-b276-192b91388aba", "address": "fa:16:3e:69:84:03", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0eb4126-40", "ovs_interfaceid": "f0eb4126-40e7-4fe8-b276-192b91388aba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.614755] env[61867]: DEBUG oslo_concurrency.lockutils [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.861s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.617692] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.328s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.618519] env[61867]: INFO nova.compute.claims [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 943.651040] env[61867]: INFO nova.scheduler.client.report [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Deleted allocations for instance 6d20f463-9198-4590-8ec2-db471c64ba7c [ 943.662554] env[61867]: DEBUG nova.compute.manager [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Received event network-changed-c05702c4-5138-450d-82b2-790b86d8a59b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.662959] env[61867]: DEBUG nova.compute.manager [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Refreshing instance network info cache due to event network-changed-c05702c4-5138-450d-82b2-790b86d8a59b. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 943.663289] env[61867]: DEBUG oslo_concurrency.lockutils [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] Acquiring lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.711168] env[61867]: DEBUG oslo_concurrency.lockutils [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.711493] env[61867]: DEBUG nova.compute.manager [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Instance network_info: |[{"id": "c05702c4-5138-450d-82b2-790b86d8a59b", "address": "fa:16:3e:cf:d4:2a", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc05702c4-51", "ovs_interfaceid": "c05702c4-5138-450d-82b2-790b86d8a59b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 943.712164] env[61867]: DEBUG oslo_concurrency.lockutils [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] Acquired lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.712379] env[61867]: DEBUG nova.network.neutron [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Refreshing network info cache for port c05702c4-5138-450d-82b2-790b86d8a59b {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 943.716027] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:d4:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '11da2092-76f7-447e-babb-8fc14ad39a71', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c05702c4-5138-450d-82b2-790b86d8a59b', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 943.721930] env[61867]: DEBUG oslo.service.loopingcall [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 943.723098] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 943.723435] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-afb25701-2d0c-4075-8f9d-eb99e3b6d8db {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.754166] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 943.754166] env[61867]: value = "task-1276928" [ 943.754166] env[61867]: _type = "Task" [ 943.754166] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.759013] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276928, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.814446] env[61867]: DEBUG nova.network.neutron [-] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.093750] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.166425] env[61867]: DEBUG oslo_concurrency.lockutils [None req-26db92b4-95cb-489d-a9c3-2575fdeb709a tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "6d20f463-9198-4590-8ec2-db471c64ba7c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.430s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.170299] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "6d20f463-9198-4590-8ec2-db471c64ba7c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 3.780s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.170299] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "6d20f463-9198-4590-8ec2-db471c64ba7c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.170299] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "6d20f463-9198-4590-8ec2-db471c64ba7c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.170299] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "6d20f463-9198-4590-8ec2-db471c64ba7c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.172058] env[61867]: INFO nova.compute.manager [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Terminating instance [ 944.173582] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "refresh_cache-6d20f463-9198-4590-8ec2-db471c64ba7c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.173744] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquired lock "refresh_cache-6d20f463-9198-4590-8ec2-db471c64ba7c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.173920] env[61867]: DEBUG nova.network.neutron [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 944.261243] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276928, 'name': CreateVM_Task, 'duration_secs': 0.410701} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.261243] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 944.261739] env[61867]: DEBUG oslo_concurrency.lockutils [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.262125] env[61867]: DEBUG oslo_concurrency.lockutils [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.262304] env[61867]: DEBUG oslo_concurrency.lockutils [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 944.262883] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6499b5f-313d-4404-9fb7-3c6d1ac3400d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.270035] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 944.270035] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f8225a-c576-606f-f076-33b32ebb3daf" [ 944.270035] env[61867]: _type = "Task" [ 944.270035] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.278285] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f8225a-c576-606f-f076-33b32ebb3daf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.322839] env[61867]: INFO nova.compute.manager [-] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Took 2.36 seconds to deallocate network for instance. [ 944.437272] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "3905193e-04da-439a-bf6c-16f638a692bf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.437596] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "3905193e-04da-439a-bf6c-16f638a692bf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.437842] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "3905193e-04da-439a-bf6c-16f638a692bf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.437972] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "3905193e-04da-439a-bf6c-16f638a692bf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.438167] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "3905193e-04da-439a-bf6c-16f638a692bf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.440391] env[61867]: INFO nova.compute.manager [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Terminating instance [ 944.442525] env[61867]: DEBUG nova.compute.manager [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 944.442808] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 944.443731] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bd7c384-d9a3-403a-a05e-d920cc54d1da {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.447536] env[61867]: DEBUG nova.network.neutron [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Updated VIF entry in instance network info cache for port c05702c4-5138-450d-82b2-790b86d8a59b. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 944.447930] env[61867]: DEBUG nova.network.neutron [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Updating instance_info_cache with network_info: [{"id": "c05702c4-5138-450d-82b2-790b86d8a59b", "address": "fa:16:3e:cf:d4:2a", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc05702c4-51", "ovs_interfaceid": "c05702c4-5138-450d-82b2-790b86d8a59b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.459246] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 944.459524] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c571b106-22de-44a1-a97d-c962d47ea8ff {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.468018] env[61867]: DEBUG oslo_vmware.api [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 944.468018] env[61867]: value = "task-1276929" [ 944.468018] env[61867]: _type = "Task" [ 944.468018] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.478432] env[61867]: DEBUG oslo_vmware.api [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276929, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.621623] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a968d9b0-9f6c-4b5a-ba5b-08033522b760 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.645872] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d70990e-6bfd-47e2-9a18-075834310e71 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.656187] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updating instance 'abb41c0c-6d0d-4147-a4af-554ab7d9e921' progress to 83 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 944.675915] env[61867]: DEBUG nova.compute.utils [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Can not refresh info_cache because instance was not found {{(pid=61867) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 944.694431] env[61867]: DEBUG nova.network.neutron [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 944.698421] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "1df8427c-e75d-4b60-a92a-b5ba76b67081" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.698748] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "1df8427c-e75d-4b60-a92a-b5ba76b67081" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.698991] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "1df8427c-e75d-4b60-a92a-b5ba76b67081-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.699244] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "1df8427c-e75d-4b60-a92a-b5ba76b67081-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.699456] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "1df8427c-e75d-4b60-a92a-b5ba76b67081-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.702216] env[61867]: INFO nova.compute.manager [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Terminating instance [ 944.706205] env[61867]: DEBUG nova.compute.manager [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 944.706205] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 944.706205] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87584d93-1231-4294-8b03-d81f586c3833 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.719682] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 944.720242] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-597cc300-06fa-4d38-8729-fd56994f89d2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.723415] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.723714] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.723929] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.724676] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.724676] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.730902] env[61867]: DEBUG oslo_vmware.api [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 944.730902] env[61867]: value = "task-1276930" [ 944.730902] env[61867]: _type = "Task" [ 944.730902] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.731443] env[61867]: INFO nova.compute.manager [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Terminating instance [ 944.733625] env[61867]: DEBUG nova.compute.manager [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 944.733871] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 944.739021] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2d8530-6925-4951-a34c-46ac371f2006 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.746952] env[61867]: DEBUG oslo_vmware.api [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276930, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.752618] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 944.753127] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-613060af-10ea-46c2-822e-cc75110cf027 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.762603] env[61867]: DEBUG oslo_vmware.api [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 944.762603] env[61867]: value = "task-1276931" [ 944.762603] env[61867]: _type = "Task" [ 944.762603] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.779166] env[61867]: DEBUG oslo_vmware.api [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276931, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.785484] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f8225a-c576-606f-f076-33b32ebb3daf, 'name': SearchDatastore_Task, 'duration_secs': 0.01761} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.790491] env[61867]: DEBUG oslo_concurrency.lockutils [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.790781] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 944.791051] env[61867]: DEBUG oslo_concurrency.lockutils [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.791208] env[61867]: DEBUG oslo_concurrency.lockutils [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.791409] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 944.791947] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-19067f41-d25f-4810-955f-5a82b0d4c304 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.803977] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 944.804281] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 944.805280] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b24c0caa-ba29-4198-ba3f-cff126c5aaa1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.815738] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 944.815738] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528f8c24-368a-da0b-0602-39a0d35057ea" [ 944.815738] env[61867]: _type = "Task" [ 944.815738] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.826299] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528f8c24-368a-da0b-0602-39a0d35057ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.833431] env[61867]: DEBUG oslo_concurrency.lockutils [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.849219] env[61867]: DEBUG nova.network.neutron [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.878578] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 944.880697] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 944.880697] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Starting heal instance info cache {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 944.880697] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Rebuilding the list of instances to heal {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 944.947244] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d27e018-f252-4bfe-ad7b-f689f4461420 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.950696] env[61867]: DEBUG oslo_concurrency.lockutils [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] Releasing lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.951090] env[61867]: DEBUG nova.compute.manager [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Received event network-vif-deleted-017138c2-47c5-4033-a0b5-42707d74a494 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.951363] env[61867]: INFO nova.compute.manager [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Neutron deleted interface 017138c2-47c5-4033-a0b5-42707d74a494; detaching it from the instance and deleting it from the info cache [ 944.951817] env[61867]: DEBUG nova.network.neutron [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Updating instance_info_cache with network_info: [{"id": "d9630611-bb16-428c-ade6-78ff0355d81d", "address": "fa:16:3e:b0:dd:b6", "network": {"id": "b49f45d7-ec84-4a83-b8af-9aac5a49ce7a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1342995777", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.73", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9630611-bb", "ovs_interfaceid": "d9630611-bb16-428c-ade6-78ff0355d81d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "74ad152e-530d-49fd-bdc1-30e366f5f9ac", "address": "fa:16:3e:3d:21:12", "network": {"id": "b49f45d7-ec84-4a83-b8af-9aac5a49ce7a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1342995777", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.148", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74ad152e-53", "ovs_interfaceid": "74ad152e-530d-49fd-bdc1-30e366f5f9ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.960706] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b48b7d-75e4-4920-8f9d-38b628b60166 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.001039] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4672e6be-09b9-4579-82e1-5a587bc0b1b3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.007640] env[61867]: DEBUG oslo_vmware.api [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276929, 'name': PowerOffVM_Task, 'duration_secs': 0.201483} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.008269] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 945.008443] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 945.008707] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c915a617-09bb-417d-a501-2c4344acf998 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.014015] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f4cfbc-4142-4d37-8386-4f4e90bc1017 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.029065] env[61867]: DEBUG nova.compute.provider_tree [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.074904] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 945.075109] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 945.075349] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Deleting the datastore file [datastore2] 3905193e-04da-439a-bf6c-16f638a692bf {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 945.075591] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2447acb7-01cc-4690-b5ad-b0ebe8804b98 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.083295] env[61867]: DEBUG oslo_vmware.api [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 945.083295] env[61867]: value = "task-1276933" [ 945.083295] env[61867]: _type = "Task" [ 945.083295] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.092876] env[61867]: DEBUG oslo_vmware.api [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276933, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.163263] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a55f6745-459f-4528-8018-32e56d5b818b tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updating instance 'abb41c0c-6d0d-4147-a4af-554ab7d9e921' progress to 100 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 945.243417] env[61867]: DEBUG oslo_vmware.api [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276930, 'name': PowerOffVM_Task, 'duration_secs': 0.26132} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.243708] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 945.243880] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 945.244143] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-08d9061c-a4c2-439f-99ad-a3c4c5857c41 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.271778] env[61867]: DEBUG oslo_vmware.api [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276931, 'name': PowerOffVM_Task, 'duration_secs': 0.198771} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.272134] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 945.272318] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 945.272566] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7055cfbd-9865-4663-890c-bea60f5f46a6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.308525] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 945.308798] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 945.308999] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Deleting the datastore file [datastore1] 1df8427c-e75d-4b60-a92a-b5ba76b67081 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 945.309304] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c75f7238-698e-4e3b-a574-a8fcf6b181b1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.317605] env[61867]: DEBUG oslo_vmware.api [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for the task: (returnval){ [ 945.317605] env[61867]: value = "task-1276936" [ 945.317605] env[61867]: _type = "Task" [ 945.317605] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.330120] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528f8c24-368a-da0b-0602-39a0d35057ea, 'name': SearchDatastore_Task, 'duration_secs': 0.022016} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.333905] env[61867]: DEBUG oslo_vmware.api [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276936, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.335136] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75651b9b-0e53-4de6-85fb-c2d5c6c92f3b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.337644] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 945.338073] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 945.338073] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleting the datastore file [datastore1] 705a32b6-67f9-42cc-b4d0-f6d1783c68b5 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 945.338379] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-344ee541-4c95-4a2c-89d2-ddbe352bd6f3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.343936] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 945.343936] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52af610c-1726-eedd-4d67-b3d2599d8efe" [ 945.343936] env[61867]: _type = "Task" [ 945.343936] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.345616] env[61867]: DEBUG oslo_concurrency.lockutils [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "816ca796-d8ed-4843-9b1f-f169f48ff047" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.345925] env[61867]: DEBUG oslo_concurrency.lockutils [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "816ca796-d8ed-4843-9b1f-f169f48ff047" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.346206] env[61867]: DEBUG oslo_concurrency.lockutils [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "816ca796-d8ed-4843-9b1f-f169f48ff047-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.346551] env[61867]: DEBUG oslo_concurrency.lockutils [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "816ca796-d8ed-4843-9b1f-f169f48ff047-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.346784] env[61867]: DEBUG oslo_concurrency.lockutils [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "816ca796-d8ed-4843-9b1f-f169f48ff047-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.348591] env[61867]: DEBUG oslo_vmware.api [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 945.348591] env[61867]: value = "task-1276937" [ 945.348591] env[61867]: _type = "Task" [ 945.348591] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.352576] env[61867]: INFO nova.compute.manager [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Terminating instance [ 945.357073] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Releasing lock "refresh_cache-6d20f463-9198-4590-8ec2-db471c64ba7c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.357484] env[61867]: DEBUG nova.compute.manager [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 945.357715] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 945.358375] env[61867]: DEBUG nova.compute.manager [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 945.358585] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 945.359265] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-209349d0-87a9-40c6-a9a9-28c35be2f578 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.361925] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc3790dc-f93a-46c8-90a0-0b4644b8d6fc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.370319] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52af610c-1726-eedd-4d67-b3d2599d8efe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.374754] env[61867]: DEBUG oslo_vmware.api [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276937, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.377655] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 945.380488] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c07f802-9b0d-4079-aead-efe25df2d767 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.390789] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1f4d417-1c2a-4b91-a6ca-a92157acd4e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.396792] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Skipping network cache update for instance because it is being deleted. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 945.396792] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Skipping network cache update for instance because it is being deleted. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 945.396792] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Skipping network cache update for instance because it is being deleted. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 945.396985] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Skipping network cache update for instance because it is being deleted. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 945.396985] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 945.397124] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 945.397334] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.397456] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquired lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.397590] env[61867]: DEBUG nova.network.neutron [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Forcefully refreshing network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 945.397753] env[61867]: DEBUG nova.objects.instance [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lazy-loading 'info_cache' on Instance uuid abb41c0c-6d0d-4147-a4af-554ab7d9e921 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.402438] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.402683] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.407573] env[61867]: DEBUG oslo_vmware.api [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 945.407573] env[61867]: value = "task-1276938" [ 945.407573] env[61867]: _type = "Task" [ 945.407573] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.417918] env[61867]: DEBUG oslo_vmware.api [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276938, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.429464] env[61867]: WARNING nova.virt.vmwareapi.vmops [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6d20f463-9198-4590-8ec2-db471c64ba7c could not be found. [ 945.429769] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 945.430054] env[61867]: INFO nova.compute.manager [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Took 0.07 seconds to destroy the instance on the hypervisor. [ 945.430361] env[61867]: DEBUG oslo.service.loopingcall [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.430709] env[61867]: DEBUG nova.compute.manager [-] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 945.430836] env[61867]: DEBUG nova.network.neutron [-] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 945.449411] env[61867]: DEBUG nova.network.neutron [-] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 945.454740] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cce8f577-8b98-4329-ab31-c3aba6bb9528 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.465583] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95245a5f-17b9-415d-8d97-1e1d9283fd7a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.498767] env[61867]: DEBUG nova.compute.manager [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Detach interface failed, port_id=017138c2-47c5-4033-a0b5-42707d74a494, reason: Instance 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 945.499029] env[61867]: DEBUG nova.compute.manager [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Received event network-vif-deleted-d9630611-bb16-428c-ade6-78ff0355d81d {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 945.499214] env[61867]: INFO nova.compute.manager [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Neutron deleted interface d9630611-bb16-428c-ade6-78ff0355d81d; detaching it from the instance and deleting it from the info cache [ 945.499473] env[61867]: DEBUG nova.network.neutron [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Updating instance_info_cache with network_info: [{"id": "74ad152e-530d-49fd-bdc1-30e366f5f9ac", "address": "fa:16:3e:3d:21:12", "network": {"id": "b49f45d7-ec84-4a83-b8af-9aac5a49ce7a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1342995777", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.148", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74ad152e-53", "ovs_interfaceid": "74ad152e-530d-49fd-bdc1-30e366f5f9ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.532133] env[61867]: DEBUG nova.scheduler.client.report [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 945.592812] env[61867]: DEBUG oslo_vmware.api [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276933, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.27925} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.593034] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 945.593220] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 945.593406] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 945.593584] env[61867]: INFO nova.compute.manager [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Took 1.15 seconds to destroy the instance on the hypervisor. [ 945.593828] env[61867]: DEBUG oslo.service.loopingcall [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.594037] env[61867]: DEBUG nova.compute.manager [-] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 945.594139] env[61867]: DEBUG nova.network.neutron [-] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 945.831900] env[61867]: DEBUG oslo_vmware.api [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Task: {'id': task-1276936, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176228} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.832929] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 945.832929] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 945.832929] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 945.832929] env[61867]: INFO nova.compute.manager [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Took 1.13 seconds to destroy the instance on the hypervisor. [ 945.834779] env[61867]: DEBUG oslo.service.loopingcall [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.834779] env[61867]: DEBUG nova.compute.manager [-] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 945.834779] env[61867]: DEBUG nova.network.neutron [-] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 945.856531] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52af610c-1726-eedd-4d67-b3d2599d8efe, 'name': SearchDatastore_Task, 'duration_secs': 0.022464} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.860068] env[61867]: DEBUG oslo_concurrency.lockutils [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.860068] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 4d7bfb31-d565-49e4-8c51-0122acebba2d/4d7bfb31-d565-49e4-8c51-0122acebba2d.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 945.860318] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cb00b70e-a956-4a45-a12f-0add77a95b2e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.864874] env[61867]: DEBUG nova.compute.manager [req-f7ed8cf8-5055-4a79-b129-9a493db337d0 req-23bded49-aa66-4823-9cec-fe89eb2038f4 service nova] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Received event network-vif-deleted-1a175a33-e511-4f6a-a971-708994d51259 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 945.865092] env[61867]: INFO nova.compute.manager [req-f7ed8cf8-5055-4a79-b129-9a493db337d0 req-23bded49-aa66-4823-9cec-fe89eb2038f4 service nova] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Neutron deleted interface 1a175a33-e511-4f6a-a971-708994d51259; detaching it from the instance and deleting it from the info cache [ 945.865278] env[61867]: DEBUG nova.network.neutron [req-f7ed8cf8-5055-4a79-b129-9a493db337d0 req-23bded49-aa66-4823-9cec-fe89eb2038f4 service nova] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.869819] env[61867]: DEBUG oslo_vmware.api [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276937, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.231785} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.870240] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 945.870429] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 945.870649] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 945.870836] env[61867]: INFO nova.compute.manager [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Took 1.14 seconds to destroy the instance on the hypervisor. [ 945.871086] env[61867]: DEBUG oslo.service.loopingcall [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.871499] env[61867]: DEBUG nova.compute.manager [-] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 945.871596] env[61867]: DEBUG nova.network.neutron [-] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 945.875170] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 945.875170] env[61867]: value = "task-1276939" [ 945.875170] env[61867]: _type = "Task" [ 945.875170] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.885849] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276939, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.904498] env[61867]: DEBUG nova.compute.manager [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 945.923443] env[61867]: DEBUG oslo_vmware.api [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276938, 'name': PowerOffVM_Task, 'duration_secs': 0.227611} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.923734] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 945.923928] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 945.924512] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-42d265b7-ba11-4e4e-baa2-1bd53ad7a6a5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.951486] env[61867]: DEBUG nova.network.neutron [-] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.994027] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 945.994342] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 945.994567] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Deleting the datastore file [datastore2] 816ca796-d8ed-4843-9b1f-f169f48ff047 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 945.994841] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-791fe40a-018c-4fcb-8a9b-824a69b49fb1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.002885] env[61867]: DEBUG oslo_vmware.api [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for the task: (returnval){ [ 946.002885] env[61867]: value = "task-1276941" [ 946.002885] env[61867]: _type = "Task" [ 946.002885] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.003155] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-28735bd5-d7b0-47af-90aa-04146cec3c8e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.017013] env[61867]: DEBUG oslo_vmware.api [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276941, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.021409] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e147d02f-30fc-4618-81fd-9dc4bc76e415 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.037338] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.420s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.038070] env[61867]: DEBUG nova.compute.manager [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 946.041162] env[61867]: DEBUG oslo_concurrency.lockutils [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.743s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.043129] env[61867]: INFO nova.compute.claims [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 946.059737] env[61867]: DEBUG nova.compute.manager [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Detach interface failed, port_id=d9630611-bb16-428c-ade6-78ff0355d81d, reason: Instance 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 946.060273] env[61867]: DEBUG nova.compute.manager [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Received event network-changed-4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.060736] env[61867]: DEBUG nova.compute.manager [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Refreshing instance network info cache due to event network-changed-4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 946.061145] env[61867]: DEBUG oslo_concurrency.lockutils [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] Acquiring lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.061370] env[61867]: DEBUG oslo_concurrency.lockutils [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] Acquired lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.061723] env[61867]: DEBUG nova.network.neutron [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Refreshing network info cache for port 4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 946.339901] env[61867]: DEBUG nova.network.neutron [-] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.372383] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-51820b51-d11b-4f18-bb58-6f03ac437607 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.396783] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276939, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.401262] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9330bb6-4325-42ea-890d-199c1dea6074 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.440454] env[61867]: DEBUG nova.compute.manager [req-f7ed8cf8-5055-4a79-b129-9a493db337d0 req-23bded49-aa66-4823-9cec-fe89eb2038f4 service nova] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Detach interface failed, port_id=1a175a33-e511-4f6a-a971-708994d51259, reason: Instance 3905193e-04da-439a-bf6c-16f638a692bf could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 946.444923] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.455934] env[61867]: INFO nova.compute.manager [-] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Took 1.02 seconds to deallocate network for instance. [ 946.517714] env[61867]: DEBUG oslo_vmware.api [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Task: {'id': task-1276941, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228587} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.518192] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.518373] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 946.518552] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 946.518723] env[61867]: INFO nova.compute.manager [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Took 1.16 seconds to destroy the instance on the hypervisor. [ 946.518955] env[61867]: DEBUG oslo.service.loopingcall [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.519228] env[61867]: DEBUG nova.compute.manager [-] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 946.519319] env[61867]: DEBUG nova.network.neutron [-] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 946.548146] env[61867]: DEBUG nova.compute.utils [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 946.549507] env[61867]: DEBUG nova.compute.manager [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 946.549674] env[61867]: DEBUG nova.network.neutron [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 946.630692] env[61867]: DEBUG nova.policy [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'adcd9eb75ecc4eccb335ebfec207a900', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7316e4f263a9432ab2f9f91484d62f58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 946.801612] env[61867]: DEBUG nova.network.neutron [-] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.842501] env[61867]: INFO nova.compute.manager [-] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Took 1.25 seconds to deallocate network for instance. [ 946.892347] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276939, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.605445} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.892674] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 4d7bfb31-d565-49e4-8c51-0122acebba2d/4d7bfb31-d565-49e4-8c51-0122acebba2d.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 946.892931] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 946.893321] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ddec48e9-67a8-4e88-98cd-cb289178ed93 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.900598] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 946.900598] env[61867]: value = "task-1276942" [ 946.900598] env[61867]: _type = "Task" [ 946.900598] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.910779] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276942, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.962757] env[61867]: INFO nova.compute.manager [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Instance disappeared during terminate [ 946.964359] env[61867]: DEBUG oslo_concurrency.lockutils [None req-21298aed-1c63-45cf-a524-0c827668d5c3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "6d20f463-9198-4590-8ec2-db471c64ba7c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.796s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.020178] env[61867]: DEBUG nova.network.neutron [-] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.053413] env[61867]: DEBUG nova.compute.manager [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 947.067259] env[61867]: DEBUG nova.network.neutron [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Updated VIF entry in instance network info cache for port 4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 947.067915] env[61867]: DEBUG nova.network.neutron [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Updating instance_info_cache with network_info: [{"id": "4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1", "address": "fa:16:3e:ce:d6:b5", "network": {"id": "ab6ec5b9-8193-43c9-a3ce-d63f31e5a001", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1573384141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35aa85105b394369b17ef4ce51a49066", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f63bd0e-c1", "ovs_interfaceid": "4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.154715] env[61867]: DEBUG nova.network.neutron [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Successfully created port: df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 947.276548] env[61867]: DEBUG nova.network.neutron [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updating instance_info_cache with network_info: [{"id": "f0eb4126-40e7-4fe8-b276-192b91388aba", "address": "fa:16:3e:69:84:03", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0eb4126-40", "ovs_interfaceid": "f0eb4126-40e7-4fe8-b276-192b91388aba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.306112] env[61867]: INFO nova.compute.manager [-] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Took 1.43 seconds to deallocate network for instance. [ 947.325613] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5070180-bf85-4be5-a600-5de51d8a7e57 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.334690] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52df4a25-0405-4127-812f-0c5584f0e26a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.366306] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.366792] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a64d5b2-dec7-4a1e-8e30-01ef148002b4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.375580] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6632661-dd39-4228-8f9f-3330a0510487 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.398082] env[61867]: DEBUG nova.compute.provider_tree [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.412510] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276942, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.195375} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.413142] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 947.413645] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cec8624-6ba1-4121-9ee0-10e4cf7926a0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.439248] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 4d7bfb31-d565-49e4-8c51-0122acebba2d/4d7bfb31-d565-49e4-8c51-0122acebba2d.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 947.439870] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c7917b6-fa92-40ce-88d3-896973fbb030 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.461986] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 947.461986] env[61867]: value = "task-1276943" [ 947.461986] env[61867]: _type = "Task" [ 947.461986] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.471231] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276943, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.523897] env[61867]: INFO nova.compute.manager [-] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Took 1.69 seconds to deallocate network for instance. [ 947.535081] env[61867]: DEBUG nova.network.neutron [-] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.576237] env[61867]: DEBUG oslo_concurrency.lockutils [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] Releasing lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.576457] env[61867]: DEBUG nova.compute.manager [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Received event network-changed-4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.576645] env[61867]: DEBUG nova.compute.manager [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Refreshing instance network info cache due to event network-changed-4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 947.576864] env[61867]: DEBUG oslo_concurrency.lockutils [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] Acquiring lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.577018] env[61867]: DEBUG oslo_concurrency.lockutils [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] Acquired lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.577197] env[61867]: DEBUG nova.network.neutron [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Refreshing network info cache for port 4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 947.780182] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Releasing lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.780395] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updated the network info_cache for instance {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 947.780661] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 947.780849] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 947.780930] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 947.781075] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 947.781223] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 947.781370] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 947.781496] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61867) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 947.781638] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 947.815655] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.896066] env[61867]: DEBUG nova.compute.manager [req-9ff7a501-b830-4536-a963-f376b4c686bb req-395803f1-77d4-4e54-a953-f0eb1c33fe2d service nova] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Received event network-vif-deleted-254a9643-f941-493e-8c87-2932a6cc00ce {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.896309] env[61867]: DEBUG nova.compute.manager [req-9ff7a501-b830-4536-a963-f376b4c686bb req-395803f1-77d4-4e54-a953-f0eb1c33fe2d service nova] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Received event network-vif-deleted-943c04a5-7eba-46db-ab4f-93f7f84be14e {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.896493] env[61867]: DEBUG nova.compute.manager [req-9ff7a501-b830-4536-a963-f376b4c686bb req-395803f1-77d4-4e54-a953-f0eb1c33fe2d service nova] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Received event network-vif-deleted-db6e6b48-df38-493c-bc63-881519b16b64 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.902043] env[61867]: DEBUG nova.scheduler.client.report [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 947.974258] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276943, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.030425] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.037089] env[61867]: INFO nova.compute.manager [-] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Took 1.52 seconds to deallocate network for instance. [ 948.061424] env[61867]: DEBUG nova.compute.manager [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 948.075816] env[61867]: DEBUG oslo_concurrency.lockutils [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.075816] env[61867]: DEBUG oslo_concurrency.lockutils [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.076016] env[61867]: DEBUG nova.compute.manager [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Going to confirm migration 3 {{(pid=61867) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 948.093313] env[61867]: DEBUG nova.virt.hardware [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 948.093619] env[61867]: DEBUG nova.virt.hardware [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 948.093788] env[61867]: DEBUG nova.virt.hardware [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 948.094022] env[61867]: DEBUG nova.virt.hardware [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 948.094193] env[61867]: DEBUG nova.virt.hardware [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 948.094364] env[61867]: DEBUG nova.virt.hardware [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 948.094610] env[61867]: DEBUG nova.virt.hardware [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 948.094797] env[61867]: DEBUG nova.virt.hardware [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 948.094990] env[61867]: DEBUG nova.virt.hardware [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 948.095210] env[61867]: DEBUG nova.virt.hardware [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 948.095487] env[61867]: DEBUG nova.virt.hardware [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 948.097979] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce50bcf-2fca-4398-b371-62ede26b8770 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.107554] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d05b837f-4ef3-4270-9043-ef3194cc2bc7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.284851] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.404119] env[61867]: DEBUG nova.network.neutron [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Updated VIF entry in instance network info cache for port 4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 948.404568] env[61867]: DEBUG nova.network.neutron [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Updating instance_info_cache with network_info: [{"id": "4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1", "address": "fa:16:3e:ce:d6:b5", "network": {"id": "ab6ec5b9-8193-43c9-a3ce-d63f31e5a001", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1573384141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35aa85105b394369b17ef4ce51a49066", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f63bd0e-c1", "ovs_interfaceid": "4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.406373] env[61867]: DEBUG oslo_concurrency.lockutils [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.406858] env[61867]: DEBUG nova.compute.manager [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 948.409529] env[61867]: DEBUG oslo_concurrency.lockutils [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.576s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.409747] env[61867]: DEBUG nova.objects.instance [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lazy-loading 'resources' on Instance uuid 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.474730] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276943, 'name': ReconfigVM_Task, 'duration_secs': 0.682938} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.475050] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 4d7bfb31-d565-49e4-8c51-0122acebba2d/4d7bfb31-d565-49e4-8c51-0122acebba2d.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 948.475707] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-577075bb-f21d-4d40-b90f-c3299a70799a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.483883] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 948.483883] env[61867]: value = "task-1276944" [ 948.483883] env[61867]: _type = "Task" [ 948.483883] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.506179] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276944, 'name': Rename_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.544194] env[61867]: DEBUG oslo_concurrency.lockutils [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.613401] env[61867]: DEBUG oslo_concurrency.lockutils [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.613596] env[61867]: DEBUG oslo_concurrency.lockutils [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.613779] env[61867]: DEBUG nova.network.neutron [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 948.613974] env[61867]: DEBUG nova.objects.instance [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lazy-loading 'info_cache' on Instance uuid abb41c0c-6d0d-4147-a4af-554ab7d9e921 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.636511] env[61867]: DEBUG nova.network.neutron [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Successfully updated port: df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 948.912015] env[61867]: DEBUG oslo_concurrency.lockutils [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] Releasing lock "refresh_cache-7479bf91-5aef-4e75-a127-7e82ae15a003" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.912015] env[61867]: DEBUG nova.compute.manager [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Received event network-vif-deleted-74ad152e-530d-49fd-bdc1-30e366f5f9ac {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.912015] env[61867]: INFO nova.compute.manager [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Neutron deleted interface 74ad152e-530d-49fd-bdc1-30e366f5f9ac; detaching it from the instance and deleting it from the info cache [ 948.912015] env[61867]: DEBUG nova.network.neutron [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.916797] env[61867]: DEBUG nova.compute.utils [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 948.918035] env[61867]: DEBUG nova.compute.manager [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 948.918035] env[61867]: DEBUG nova.network.neutron [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 948.957331] env[61867]: DEBUG nova.policy [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a013ca682054d16a6702230a22c180f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c28a8e133444440eb83dc6a848ef591d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 948.993621] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276944, 'name': Rename_Task, 'duration_secs': 0.154981} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.993933] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 948.994234] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6478b9b3-2af4-485d-a5a1-0e6df488fd8b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.001743] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 949.001743] env[61867]: value = "task-1276945" [ 949.001743] env[61867]: _type = "Task" [ 949.001743] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.010132] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276945, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.131876] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea9810d6-131e-421a-9916-e3f0e5b87d7e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.138944] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "refresh_cache-43dc6520-5256-4b5f-a273-6c9e0e407c72" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.139490] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "refresh_cache-43dc6520-5256-4b5f-a273-6c9e0e407c72" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.139696] env[61867]: DEBUG nova.network.neutron [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 949.144469] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e07c754e-37a8-4cc1-8149-d5a4eba297f8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.182972] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-537cadc3-3b6d-4553-92bf-370080ba5d82 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.192243] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e82c660-6554-465c-98a9-4a9906c390de {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.208568] env[61867]: DEBUG nova.compute.provider_tree [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.246291] env[61867]: DEBUG nova.network.neutron [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Successfully created port: a64da4a0-c1d4-4c3a-8386-a39668fa32e9 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 949.418851] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-52226e4d-5b94-4ab1-aeea-bcc6598f6369 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.421951] env[61867]: DEBUG nova.compute.manager [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 949.433236] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-676278a0-4b42-47d1-8718-874629e8f57e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.469622] env[61867]: DEBUG nova.compute.manager [req-eb8f7225-c244-4b53-92a6-572162dedf64 req-f4a05066-b96f-43ef-8151-3b49d844e82e service nova] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Detach interface failed, port_id=74ad152e-530d-49fd-bdc1-30e366f5f9ac, reason: Instance 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 949.517904] env[61867]: DEBUG oslo_vmware.api [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276945, 'name': PowerOnVM_Task, 'duration_secs': 0.515549} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.518264] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 949.518474] env[61867]: INFO nova.compute.manager [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Took 8.23 seconds to spawn the instance on the hypervisor. [ 949.518648] env[61867]: DEBUG nova.compute.manager [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 949.519460] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4433eb-6395-40c8-b9ff-aa2f35ea0daa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.674197] env[61867]: DEBUG nova.network.neutron [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 949.714162] env[61867]: DEBUG nova.scheduler.client.report [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.810289] env[61867]: DEBUG nova.network.neutron [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Updating instance_info_cache with network_info: [{"id": "df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83", "address": "fa:16:3e:7d:a5:8b", "network": {"id": "14cee476-296a-4093-99ba-c2a5ef63f7cf", "bridge": "br-int", "label": "tempest-ServersTestJSON-243183479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7316e4f263a9432ab2f9f91484d62f58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf8e3f7f-3b", "ovs_interfaceid": "df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.882407] env[61867]: DEBUG nova.network.neutron [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updating instance_info_cache with network_info: [{"id": "f0eb4126-40e7-4fe8-b276-192b91388aba", "address": "fa:16:3e:69:84:03", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0eb4126-40", "ovs_interfaceid": "f0eb4126-40e7-4fe8-b276-192b91388aba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.921863] env[61867]: DEBUG nova.compute.manager [req-682e2cf9-88f1-4d4c-a812-22c916e751d4 req-2d83d655-87d3-4941-aa67-487eb9e31bd8 service nova] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Received event network-vif-plugged-df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.922142] env[61867]: DEBUG oslo_concurrency.lockutils [req-682e2cf9-88f1-4d4c-a812-22c916e751d4 req-2d83d655-87d3-4941-aa67-487eb9e31bd8 service nova] Acquiring lock "43dc6520-5256-4b5f-a273-6c9e0e407c72-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.922395] env[61867]: DEBUG oslo_concurrency.lockutils [req-682e2cf9-88f1-4d4c-a812-22c916e751d4 req-2d83d655-87d3-4941-aa67-487eb9e31bd8 service nova] Lock "43dc6520-5256-4b5f-a273-6c9e0e407c72-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.922543] env[61867]: DEBUG oslo_concurrency.lockutils [req-682e2cf9-88f1-4d4c-a812-22c916e751d4 req-2d83d655-87d3-4941-aa67-487eb9e31bd8 service nova] Lock "43dc6520-5256-4b5f-a273-6c9e0e407c72-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.922718] env[61867]: DEBUG nova.compute.manager [req-682e2cf9-88f1-4d4c-a812-22c916e751d4 req-2d83d655-87d3-4941-aa67-487eb9e31bd8 service nova] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] No waiting events found dispatching network-vif-plugged-df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 949.922886] env[61867]: WARNING nova.compute.manager [req-682e2cf9-88f1-4d4c-a812-22c916e751d4 req-2d83d655-87d3-4941-aa67-487eb9e31bd8 service nova] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Received unexpected event network-vif-plugged-df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83 for instance with vm_state building and task_state spawning. [ 949.923086] env[61867]: DEBUG nova.compute.manager [req-682e2cf9-88f1-4d4c-a812-22c916e751d4 req-2d83d655-87d3-4941-aa67-487eb9e31bd8 service nova] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Received event network-changed-df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.923256] env[61867]: DEBUG nova.compute.manager [req-682e2cf9-88f1-4d4c-a812-22c916e751d4 req-2d83d655-87d3-4941-aa67-487eb9e31bd8 service nova] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Refreshing instance network info cache due to event network-changed-df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 949.923425] env[61867]: DEBUG oslo_concurrency.lockutils [req-682e2cf9-88f1-4d4c-a812-22c916e751d4 req-2d83d655-87d3-4941-aa67-487eb9e31bd8 service nova] Acquiring lock "refresh_cache-43dc6520-5256-4b5f-a273-6c9e0e407c72" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.040091] env[61867]: INFO nova.compute.manager [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Took 17.75 seconds to build instance. [ 950.220485] env[61867]: DEBUG oslo_concurrency.lockutils [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.810s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.222110] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.780s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.224205] env[61867]: INFO nova.compute.claims [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.246719] env[61867]: INFO nova.scheduler.client.report [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Deleted allocations for instance 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f [ 950.313104] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "refresh_cache-43dc6520-5256-4b5f-a273-6c9e0e407c72" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.313104] env[61867]: DEBUG nova.compute.manager [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Instance network_info: |[{"id": "df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83", "address": "fa:16:3e:7d:a5:8b", "network": {"id": "14cee476-296a-4093-99ba-c2a5ef63f7cf", "bridge": "br-int", "label": "tempest-ServersTestJSON-243183479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7316e4f263a9432ab2f9f91484d62f58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf8e3f7f-3b", "ovs_interfaceid": "df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 950.313104] env[61867]: DEBUG oslo_concurrency.lockutils [req-682e2cf9-88f1-4d4c-a812-22c916e751d4 req-2d83d655-87d3-4941-aa67-487eb9e31bd8 service nova] Acquired lock "refresh_cache-43dc6520-5256-4b5f-a273-6c9e0e407c72" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.313391] env[61867]: DEBUG nova.network.neutron [req-682e2cf9-88f1-4d4c-a812-22c916e751d4 req-2d83d655-87d3-4941-aa67-487eb9e31bd8 service nova] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Refreshing network info cache for port df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 950.314485] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:a5:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '085fb0ff-9285-4f1d-a008-a14da4844357', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 950.322137] env[61867]: DEBUG oslo.service.loopingcall [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.324733] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 950.325221] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-99fa0100-a066-4729-8388-183a90ed8b38 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.345866] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 950.345866] env[61867]: value = "task-1276946" [ 950.345866] env[61867]: _type = "Task" [ 950.345866] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.354107] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276946, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.385469] env[61867]: DEBUG oslo_concurrency.lockutils [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.385751] env[61867]: DEBUG nova.objects.instance [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lazy-loading 'migration_context' on Instance uuid abb41c0c-6d0d-4147-a4af-554ab7d9e921 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.434632] env[61867]: DEBUG nova.compute.manager [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 950.460974] env[61867]: DEBUG nova.virt.hardware [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 950.461264] env[61867]: DEBUG nova.virt.hardware [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 950.461463] env[61867]: DEBUG nova.virt.hardware [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 950.461681] env[61867]: DEBUG nova.virt.hardware [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 950.461854] env[61867]: DEBUG nova.virt.hardware [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 950.462158] env[61867]: DEBUG nova.virt.hardware [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 950.462410] env[61867]: DEBUG nova.virt.hardware [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 950.462597] env[61867]: DEBUG nova.virt.hardware [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 950.462775] env[61867]: DEBUG nova.virt.hardware [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 950.462975] env[61867]: DEBUG nova.virt.hardware [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 950.463202] env[61867]: DEBUG nova.virt.hardware [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.464416] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8811535a-223d-4760-8cff-4fde037a1210 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.473179] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b7128e0-a634-4601-93b1-bc4584bcb19d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.541784] env[61867]: DEBUG oslo_concurrency.lockutils [None req-908714d0-1e22-4284-b7af-05c881cbee51 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "4d7bfb31-d565-49e4-8c51-0122acebba2d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.259s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.605928] env[61867]: DEBUG nova.network.neutron [req-682e2cf9-88f1-4d4c-a812-22c916e751d4 req-2d83d655-87d3-4941-aa67-487eb9e31bd8 service nova] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Updated VIF entry in instance network info cache for port df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 950.606500] env[61867]: DEBUG nova.network.neutron [req-682e2cf9-88f1-4d4c-a812-22c916e751d4 req-2d83d655-87d3-4941-aa67-487eb9e31bd8 service nova] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Updating instance_info_cache with network_info: [{"id": "df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83", "address": "fa:16:3e:7d:a5:8b", "network": {"id": "14cee476-296a-4093-99ba-c2a5ef63f7cf", "bridge": "br-int", "label": "tempest-ServersTestJSON-243183479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7316e4f263a9432ab2f9f91484d62f58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf8e3f7f-3b", "ovs_interfaceid": "df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.756342] env[61867]: DEBUG oslo_concurrency.lockutils [None req-90680b03-21f1-4a77-92e3-613e02f83ea0 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "38edb89a-28e4-4dd2-a8aa-35cb95a72e0f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.019s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.818888] env[61867]: DEBUG nova.network.neutron [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Successfully updated port: a64da4a0-c1d4-4c3a-8386-a39668fa32e9 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 950.858283] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276946, 'name': CreateVM_Task, 'duration_secs': 0.310624} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.858580] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 950.859546] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.859839] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.860281] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 950.860920] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d9bfcf0-4fcd-4438-94b3-2b1f0009782a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.866689] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 950.866689] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e57c7e-b422-5c82-f53f-5a7f34d59076" [ 950.866689] env[61867]: _type = "Task" [ 950.866689] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.877412] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e57c7e-b422-5c82-f53f-5a7f34d59076, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.888493] env[61867]: DEBUG nova.objects.base [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61867) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 950.889623] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df30f32b-ed5a-4d45-bb90-abb1a60c59ff {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.909466] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b86104bc-ab8e-4b12-b9eb-b7eb1ea6c5c8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.916325] env[61867]: DEBUG oslo_vmware.api [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 950.916325] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52802ef2-bd5a-343b-7f9c-34ea280a34dc" [ 950.916325] env[61867]: _type = "Task" [ 950.916325] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.927588] env[61867]: DEBUG oslo_vmware.api [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52802ef2-bd5a-343b-7f9c-34ea280a34dc, 'name': SearchDatastore_Task, 'duration_secs': 0.008106} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.927588] env[61867]: DEBUG oslo_concurrency.lockutils [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.109549] env[61867]: DEBUG oslo_concurrency.lockutils [req-682e2cf9-88f1-4d4c-a812-22c916e751d4 req-2d83d655-87d3-4941-aa67-487eb9e31bd8 service nova] Releasing lock "refresh_cache-43dc6520-5256-4b5f-a273-6c9e0e407c72" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.321109] env[61867]: DEBUG oslo_concurrency.lockutils [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "refresh_cache-5cf15593-54f6-405c-8435-2e3b378983b7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.321109] env[61867]: DEBUG oslo_concurrency.lockutils [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "refresh_cache-5cf15593-54f6-405c-8435-2e3b378983b7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.321109] env[61867]: DEBUG nova.network.neutron [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 951.388034] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e57c7e-b422-5c82-f53f-5a7f34d59076, 'name': SearchDatastore_Task, 'duration_secs': 0.011071} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.388034] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.388034] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 951.388034] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.388034] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.388034] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 951.388034] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b74c300e-7ecc-47df-bc23-262587c7100e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.400228] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 951.400228] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 951.400228] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f76efd6-934c-4478-bbda-18c52e576d97 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.411018] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 951.411018] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521e8840-ec3c-6119-56af-c116221403d7" [ 951.411018] env[61867]: _type = "Task" [ 951.411018] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.417336] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521e8840-ec3c-6119-56af-c116221403d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.476017] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de8aa52-ac2d-47f3-bc85-2c4daf3e18d7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.483751] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0606169f-f9c8-4660-b57a-e5c45760664f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.517059] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d967c1b-788e-4277-81ea-3373d44be827 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.520835] env[61867]: DEBUG nova.compute.manager [req-df6c0a06-66fe-4523-b374-ff2587b34e45 req-5013c4da-7e11-42c8-9a68-04b14ee8d76d service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Received event network-changed-c05702c4-5138-450d-82b2-790b86d8a59b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.521163] env[61867]: DEBUG nova.compute.manager [req-df6c0a06-66fe-4523-b374-ff2587b34e45 req-5013c4da-7e11-42c8-9a68-04b14ee8d76d service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Refreshing instance network info cache due to event network-changed-c05702c4-5138-450d-82b2-790b86d8a59b. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 951.521497] env[61867]: DEBUG oslo_concurrency.lockutils [req-df6c0a06-66fe-4523-b374-ff2587b34e45 req-5013c4da-7e11-42c8-9a68-04b14ee8d76d service nova] Acquiring lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.521761] env[61867]: DEBUG oslo_concurrency.lockutils [req-df6c0a06-66fe-4523-b374-ff2587b34e45 req-5013c4da-7e11-42c8-9a68-04b14ee8d76d service nova] Acquired lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.522092] env[61867]: DEBUG nova.network.neutron [req-df6c0a06-66fe-4523-b374-ff2587b34e45 req-5013c4da-7e11-42c8-9a68-04b14ee8d76d service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Refreshing network info cache for port c05702c4-5138-450d-82b2-790b86d8a59b {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 951.535018] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9df5c87-32f7-4a8c-8962-6bb5e32c5419 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.546809] env[61867]: DEBUG nova.compute.provider_tree [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 951.862681] env[61867]: DEBUG nova.network.neutron [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 951.935414] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521e8840-ec3c-6119-56af-c116221403d7, 'name': SearchDatastore_Task, 'duration_secs': 0.009821} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.938250] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b01d5220-1cf1-41fe-b246-6376534862e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.945496] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 951.945496] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5284595a-70eb-dbbc-fe33-033c690f828e" [ 951.945496] env[61867]: _type = "Task" [ 951.945496] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.953915] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5284595a-70eb-dbbc-fe33-033c690f828e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.049928] env[61867]: DEBUG nova.scheduler.client.report [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.094485] env[61867]: DEBUG nova.network.neutron [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Updating instance_info_cache with network_info: [{"id": "a64da4a0-c1d4-4c3a-8386-a39668fa32e9", "address": "fa:16:3e:55:f5:77", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa64da4a0-c1", "ovs_interfaceid": "a64da4a0-c1d4-4c3a-8386-a39668fa32e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.144174] env[61867]: DEBUG nova.compute.manager [req-c351fc76-30d1-4fb4-aec2-8a689ec46879 req-854872b6-30b2-46b0-932f-fb6246a2ce1a service nova] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Received event network-vif-plugged-a64da4a0-c1d4-4c3a-8386-a39668fa32e9 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.144174] env[61867]: DEBUG oslo_concurrency.lockutils [req-c351fc76-30d1-4fb4-aec2-8a689ec46879 req-854872b6-30b2-46b0-932f-fb6246a2ce1a service nova] Acquiring lock "5cf15593-54f6-405c-8435-2e3b378983b7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.144174] env[61867]: DEBUG oslo_concurrency.lockutils [req-c351fc76-30d1-4fb4-aec2-8a689ec46879 req-854872b6-30b2-46b0-932f-fb6246a2ce1a service nova] Lock "5cf15593-54f6-405c-8435-2e3b378983b7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.149254] env[61867]: DEBUG oslo_concurrency.lockutils [req-c351fc76-30d1-4fb4-aec2-8a689ec46879 req-854872b6-30b2-46b0-932f-fb6246a2ce1a service nova] Lock "5cf15593-54f6-405c-8435-2e3b378983b7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.004s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.149254] env[61867]: DEBUG nova.compute.manager [req-c351fc76-30d1-4fb4-aec2-8a689ec46879 req-854872b6-30b2-46b0-932f-fb6246a2ce1a service nova] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] No waiting events found dispatching network-vif-plugged-a64da4a0-c1d4-4c3a-8386-a39668fa32e9 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 952.149254] env[61867]: WARNING nova.compute.manager [req-c351fc76-30d1-4fb4-aec2-8a689ec46879 req-854872b6-30b2-46b0-932f-fb6246a2ce1a service nova] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Received unexpected event network-vif-plugged-a64da4a0-c1d4-4c3a-8386-a39668fa32e9 for instance with vm_state building and task_state spawning. [ 952.149254] env[61867]: DEBUG nova.compute.manager [req-c351fc76-30d1-4fb4-aec2-8a689ec46879 req-854872b6-30b2-46b0-932f-fb6246a2ce1a service nova] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Received event network-changed-a64da4a0-c1d4-4c3a-8386-a39668fa32e9 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.149254] env[61867]: DEBUG nova.compute.manager [req-c351fc76-30d1-4fb4-aec2-8a689ec46879 req-854872b6-30b2-46b0-932f-fb6246a2ce1a service nova] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Refreshing instance network info cache due to event network-changed-a64da4a0-c1d4-4c3a-8386-a39668fa32e9. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 952.149254] env[61867]: DEBUG oslo_concurrency.lockutils [req-c351fc76-30d1-4fb4-aec2-8a689ec46879 req-854872b6-30b2-46b0-932f-fb6246a2ce1a service nova] Acquiring lock "refresh_cache-5cf15593-54f6-405c-8435-2e3b378983b7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.456860] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5284595a-70eb-dbbc-fe33-033c690f828e, 'name': SearchDatastore_Task, 'duration_secs': 0.01109} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.456860] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.457087] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 43dc6520-5256-4b5f-a273-6c9e0e407c72/43dc6520-5256-4b5f-a273-6c9e0e407c72.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 952.457364] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-29178dc4-686f-4bfa-9526-c7590c61512c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.464190] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 952.464190] env[61867]: value = "task-1276947" [ 952.464190] env[61867]: _type = "Task" [ 952.464190] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.472192] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276947, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.504212] env[61867]: DEBUG nova.network.neutron [req-df6c0a06-66fe-4523-b374-ff2587b34e45 req-5013c4da-7e11-42c8-9a68-04b14ee8d76d service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Updated VIF entry in instance network info cache for port c05702c4-5138-450d-82b2-790b86d8a59b. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 952.504617] env[61867]: DEBUG nova.network.neutron [req-df6c0a06-66fe-4523-b374-ff2587b34e45 req-5013c4da-7e11-42c8-9a68-04b14ee8d76d service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Updating instance_info_cache with network_info: [{"id": "c05702c4-5138-450d-82b2-790b86d8a59b", "address": "fa:16:3e:cf:d4:2a", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc05702c4-51", "ovs_interfaceid": "c05702c4-5138-450d-82b2-790b86d8a59b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.554893] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.555502] env[61867]: DEBUG nova.compute.manager [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 952.558350] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.193s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.558597] env[61867]: DEBUG nova.objects.instance [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lazy-loading 'resources' on Instance uuid 3905193e-04da-439a-bf6c-16f638a692bf {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.598610] env[61867]: DEBUG oslo_concurrency.lockutils [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "refresh_cache-5cf15593-54f6-405c-8435-2e3b378983b7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.598610] env[61867]: DEBUG nova.compute.manager [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Instance network_info: |[{"id": "a64da4a0-c1d4-4c3a-8386-a39668fa32e9", "address": "fa:16:3e:55:f5:77", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa64da4a0-c1", "ovs_interfaceid": "a64da4a0-c1d4-4c3a-8386-a39668fa32e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 952.598610] env[61867]: DEBUG oslo_concurrency.lockutils [req-c351fc76-30d1-4fb4-aec2-8a689ec46879 req-854872b6-30b2-46b0-932f-fb6246a2ce1a service nova] Acquired lock "refresh_cache-5cf15593-54f6-405c-8435-2e3b378983b7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.598610] env[61867]: DEBUG nova.network.neutron [req-c351fc76-30d1-4fb4-aec2-8a689ec46879 req-854872b6-30b2-46b0-932f-fb6246a2ce1a service nova] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Refreshing network info cache for port a64da4a0-c1d4-4c3a-8386-a39668fa32e9 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 952.600248] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:f5:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74e6f6e0-95e6-4531-99e9-0e78350fb655', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a64da4a0-c1d4-4c3a-8386-a39668fa32e9', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 952.608205] env[61867]: DEBUG oslo.service.loopingcall [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.610179] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 952.610179] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b29a3ccf-225e-4219-8387-143272aa0516 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.631745] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 952.631745] env[61867]: value = "task-1276948" [ 952.631745] env[61867]: _type = "Task" [ 952.631745] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.642353] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276948, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.693548] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "0ae83268-ef50-42e6-954d-5e67c4afc20e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.693751] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "0ae83268-ef50-42e6-954d-5e67c4afc20e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.978231] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276947, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.007589] env[61867]: DEBUG oslo_concurrency.lockutils [req-df6c0a06-66fe-4523-b374-ff2587b34e45 req-5013c4da-7e11-42c8-9a68-04b14ee8d76d service nova] Releasing lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.061445] env[61867]: DEBUG nova.compute.utils [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 953.063080] env[61867]: DEBUG nova.compute.manager [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 953.063288] env[61867]: DEBUG nova.network.neutron [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 953.141772] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276948, 'name': CreateVM_Task, 'duration_secs': 0.406073} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.144349] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 953.145700] env[61867]: DEBUG oslo_concurrency.lockutils [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.145920] env[61867]: DEBUG oslo_concurrency.lockutils [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.146261] env[61867]: DEBUG oslo_concurrency.lockutils [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 953.148357] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7bef425-a20b-402e-84f6-7d4c3ca26f0f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.153821] env[61867]: DEBUG nova.policy [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12ed008ec3204102b2b08e61ed24f418', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '47c5b5b783ea445eb1c02fb728b1fff2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 953.160888] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 953.160888] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52eaf78d-4085-0eb9-1c97-8c424de8ebf0" [ 953.160888] env[61867]: _type = "Task" [ 953.160888] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.177016] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52eaf78d-4085-0eb9-1c97-8c424de8ebf0, 'name': SearchDatastore_Task, 'duration_secs': 0.015873} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.179785] env[61867]: DEBUG oslo_concurrency.lockutils [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.180748] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 953.181073] env[61867]: DEBUG oslo_concurrency.lockutils [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.181275] env[61867]: DEBUG oslo_concurrency.lockutils [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.181482] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.182088] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5a6a55d2-1f73-4e7b-bbfa-aa967a7f42c1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.195433] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.195638] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 953.196855] env[61867]: DEBUG nova.compute.manager [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 953.202173] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd2f4659-963a-4e66-aa87-24a3fe1d7143 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.207084] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 953.207084] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521f68ca-0e06-9b7a-8cf7-beae819e1632" [ 953.207084] env[61867]: _type = "Task" [ 953.207084] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.218852] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521f68ca-0e06-9b7a-8cf7-beae819e1632, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.335353] env[61867]: DEBUG nova.network.neutron [req-c351fc76-30d1-4fb4-aec2-8a689ec46879 req-854872b6-30b2-46b0-932f-fb6246a2ce1a service nova] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Updated VIF entry in instance network info cache for port a64da4a0-c1d4-4c3a-8386-a39668fa32e9. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 953.335353] env[61867]: DEBUG nova.network.neutron [req-c351fc76-30d1-4fb4-aec2-8a689ec46879 req-854872b6-30b2-46b0-932f-fb6246a2ce1a service nova] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Updating instance_info_cache with network_info: [{"id": "a64da4a0-c1d4-4c3a-8386-a39668fa32e9", "address": "fa:16:3e:55:f5:77", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa64da4a0-c1", "ovs_interfaceid": "a64da4a0-c1d4-4c3a-8386-a39668fa32e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.345054] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c5eba3-4707-4bd7-9703-d8d3716625d1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.352909] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f4381e6-29a5-47af-95af-4ec09da64a15 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.388157] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8cfdf0-6ba8-4954-a181-078b08158ebe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.396354] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b89c10-f5c6-481f-8fc4-8262918f03c0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.410365] env[61867]: DEBUG nova.compute.provider_tree [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.479157] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276947, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516979} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.480681] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 43dc6520-5256-4b5f-a273-6c9e0e407c72/43dc6520-5256-4b5f-a273-6c9e0e407c72.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 953.481102] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 953.481502] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0895e387-23a0-49df-a85e-ff7f0952743a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.506123] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 953.506123] env[61867]: value = "task-1276949" [ 953.506123] env[61867]: _type = "Task" [ 953.506123] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.514723] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276949, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.569144] env[61867]: DEBUG nova.compute.manager [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 953.661121] env[61867]: DEBUG nova.network.neutron [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Successfully created port: e070a089-e0e7-4b11-9688-26d528b37e6a {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 953.723022] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]521f68ca-0e06-9b7a-8cf7-beae819e1632, 'name': SearchDatastore_Task, 'duration_secs': 0.042336} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.723022] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e624137-0a68-49bb-b554-f192d5a20d76 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.725630] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.728162] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 953.728162] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e436a8-c742-eb1e-a1fa-e770c45b4835" [ 953.728162] env[61867]: _type = "Task" [ 953.728162] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.738047] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e436a8-c742-eb1e-a1fa-e770c45b4835, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.837844] env[61867]: DEBUG oslo_concurrency.lockutils [req-c351fc76-30d1-4fb4-aec2-8a689ec46879 req-854872b6-30b2-46b0-932f-fb6246a2ce1a service nova] Releasing lock "refresh_cache-5cf15593-54f6-405c-8435-2e3b378983b7" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.913639] env[61867]: DEBUG nova.scheduler.client.report [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.016796] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276949, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.330093} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.017405] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 954.018425] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-877ae8eb-78b2-4e6c-baf5-9966f1fc27f9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.041477] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 43dc6520-5256-4b5f-a273-6c9e0e407c72/43dc6520-5256-4b5f-a273-6c9e0e407c72.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 954.041888] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14947d40-9abb-48cf-b720-5a3823375d5a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.063573] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 954.063573] env[61867]: value = "task-1276950" [ 954.063573] env[61867]: _type = "Task" [ 954.063573] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.077681] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276950, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.215539] env[61867]: DEBUG oslo_concurrency.lockutils [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquiring lock "f426dd3e-2806-4d70-8784-75c611fd03e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.215539] env[61867]: DEBUG oslo_concurrency.lockutils [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "f426dd3e-2806-4d70-8784-75c611fd03e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.239708] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e436a8-c742-eb1e-a1fa-e770c45b4835, 'name': SearchDatastore_Task, 'duration_secs': 0.05865} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.240388] env[61867]: DEBUG oslo_concurrency.lockutils [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.240682] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 5cf15593-54f6-405c-8435-2e3b378983b7/5cf15593-54f6-405c-8435-2e3b378983b7.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 954.240969] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d5fca483-9bb7-432b-a6c2-abd811280966 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.248822] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 954.248822] env[61867]: value = "task-1276951" [ 954.248822] env[61867]: _type = "Task" [ 954.248822] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.258378] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276951, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.418728] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.860s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.421597] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.606s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.421597] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.423621] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.393s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.423871] env[61867]: DEBUG nova.objects.instance [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lazy-loading 'resources' on Instance uuid 1df8427c-e75d-4b60-a92a-b5ba76b67081 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.452176] env[61867]: INFO nova.scheduler.client.report [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Deleted allocations for instance 3905193e-04da-439a-bf6c-16f638a692bf [ 954.457268] env[61867]: INFO nova.scheduler.client.report [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleted allocations for instance 705a32b6-67f9-42cc-b4d0-f6d1783c68b5 [ 954.574857] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276950, 'name': ReconfigVM_Task, 'duration_secs': 0.38109} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.575253] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 43dc6520-5256-4b5f-a273-6c9e0e407c72/43dc6520-5256-4b5f-a273-6c9e0e407c72.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 954.575827] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3fbe2d53-ae86-42c3-89dc-1781e2004d03 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.578451] env[61867]: DEBUG nova.compute.manager [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 954.587361] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 954.587361] env[61867]: value = "task-1276952" [ 954.587361] env[61867]: _type = "Task" [ 954.587361] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.597614] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276952, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.783332] env[61867]: DEBUG nova.virt.hardware [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 954.783332] env[61867]: DEBUG nova.virt.hardware [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 954.783332] env[61867]: DEBUG nova.virt.hardware [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 954.783332] env[61867]: DEBUG nova.virt.hardware [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 954.783332] env[61867]: DEBUG nova.virt.hardware [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 954.783332] env[61867]: DEBUG nova.virt.hardware [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 954.783332] env[61867]: DEBUG nova.virt.hardware [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 954.783332] env[61867]: DEBUG nova.virt.hardware [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 954.783332] env[61867]: DEBUG nova.virt.hardware [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 954.783332] env[61867]: DEBUG nova.virt.hardware [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 954.783332] env[61867]: DEBUG nova.virt.hardware [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 954.783332] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba7b33b-611b-434e-b310-da1cdb030d71 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.783332] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1171deb5-bd5a-4d4b-96fe-2a7ebc65cba1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.783332] env[61867]: DEBUG nova.compute.manager [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 954.783332] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276951, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446982} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.787814] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 5cf15593-54f6-405c-8435-2e3b378983b7/5cf15593-54f6-405c-8435-2e3b378983b7.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 954.787814] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 954.787814] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2adce48f-1bed-44fb-8278-f49f55e1b78a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.787814] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 954.787814] env[61867]: value = "task-1276953" [ 954.787814] env[61867]: _type = "Task" [ 954.787814] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.789789] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276953, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.960289] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a72f5868-8476-49fd-a57b-2a4771155c0f tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "3905193e-04da-439a-bf6c-16f638a692bf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.522s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.963722] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8f5883f6-58ce-4b4f-98f8-fe03324304b4 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "705a32b6-67f9-42cc-b4d0-f6d1783c68b5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.240s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.105139] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276952, 'name': Rename_Task, 'duration_secs': 0.212116} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.105375] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 955.105685] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-89849076-a803-423a-98ff-d13d4de961de {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.113506] env[61867]: DEBUG nova.compute.manager [req-818a3816-fc73-475a-937b-55f5269acb9c req-c997615b-3988-4dc6-b9bd-8a4008547bc2 service nova] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Received event network-vif-plugged-e070a089-e0e7-4b11-9688-26d528b37e6a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.113699] env[61867]: DEBUG oslo_concurrency.lockutils [req-818a3816-fc73-475a-937b-55f5269acb9c req-c997615b-3988-4dc6-b9bd-8a4008547bc2 service nova] Acquiring lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.113911] env[61867]: DEBUG oslo_concurrency.lockutils [req-818a3816-fc73-475a-937b-55f5269acb9c req-c997615b-3988-4dc6-b9bd-8a4008547bc2 service nova] Lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.114098] env[61867]: DEBUG oslo_concurrency.lockutils [req-818a3816-fc73-475a-937b-55f5269acb9c req-c997615b-3988-4dc6-b9bd-8a4008547bc2 service nova] Lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.114272] env[61867]: DEBUG nova.compute.manager [req-818a3816-fc73-475a-937b-55f5269acb9c req-c997615b-3988-4dc6-b9bd-8a4008547bc2 service nova] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] No waiting events found dispatching network-vif-plugged-e070a089-e0e7-4b11-9688-26d528b37e6a {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 955.114441] env[61867]: WARNING nova.compute.manager [req-818a3816-fc73-475a-937b-55f5269acb9c req-c997615b-3988-4dc6-b9bd-8a4008547bc2 service nova] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Received unexpected event network-vif-plugged-e070a089-e0e7-4b11-9688-26d528b37e6a for instance with vm_state building and task_state spawning. [ 955.116195] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 955.116195] env[61867]: value = "task-1276954" [ 955.116195] env[61867]: _type = "Task" [ 955.116195] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.127379] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276954, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.137873] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d2e2f68-a428-49cd-b4fe-37e20e6f7dff {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.145801] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97cd553d-b5e8-4b26-9e98-6a70f3e4cc1e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.179249] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75c99ed2-30db-4c3c-a949-86ed3b67d72e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.188309] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b9cd9e4-7a2a-45c3-bcc0-20d77124ab88 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.202791] env[61867]: DEBUG nova.compute.provider_tree [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.211296] env[61867]: DEBUG nova.network.neutron [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Successfully updated port: e070a089-e0e7-4b11-9688-26d528b37e6a {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 955.244797] env[61867]: DEBUG oslo_concurrency.lockutils [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.287915] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276953, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065499} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.288227] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 955.289010] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214dcf3f-6282-4be0-adcf-9ccb48b13c62 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.311225] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 5cf15593-54f6-405c-8435-2e3b378983b7/5cf15593-54f6-405c-8435-2e3b378983b7.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.311504] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3b2e4a8-cf4c-4e49-82a5-447117f5f4e3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.330970] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 955.330970] env[61867]: value = "task-1276955" [ 955.330970] env[61867]: _type = "Task" [ 955.330970] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.339272] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276955, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.626871] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276954, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.705928] env[61867]: DEBUG nova.scheduler.client.report [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 955.715290] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "refresh_cache-6e2136e7-c7e0-4a98-9899-f79c10f0e703" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.715290] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "refresh_cache-6e2136e7-c7e0-4a98-9899-f79c10f0e703" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.715290] env[61867]: DEBUG nova.network.neutron [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 955.844107] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276955, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.127287] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276954, 'name': PowerOnVM_Task} progress is 81%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.211861] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.788s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.214807] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 7.930s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.214807] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.214807] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61867) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 956.215546] env[61867]: DEBUG oslo_concurrency.lockutils [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.671s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.215546] env[61867]: DEBUG nova.objects.instance [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lazy-loading 'resources' on Instance uuid 816ca796-d8ed-4843-9b1f-f169f48ff047 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.216819] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2271521-19fc-482a-8ced-4a0282a54969 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.228033] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d0e669-fbdb-47e7-9d8a-78e925d40a5e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.243667] env[61867]: INFO nova.scheduler.client.report [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Deleted allocations for instance 1df8427c-e75d-4b60-a92a-b5ba76b67081 [ 956.248341] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c1260a8-011b-4a06-a0a6-d9ed60b27256 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.251543] env[61867]: DEBUG nova.network.neutron [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 956.259071] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df99d34f-a292-4719-acbe-13a553f2568d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.291600] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179481MB free_disk=141GB free_vcpus=48 pci_devices=None {{(pid=61867) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 956.291782] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.343032] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276955, 'name': ReconfigVM_Task, 'duration_secs': 0.774585} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.343566] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 5cf15593-54f6-405c-8435-2e3b378983b7/5cf15593-54f6-405c-8435-2e3b378983b7.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.344339] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-305e945a-287a-4d21-b656-9f34997835e5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.353045] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 956.353045] env[61867]: value = "task-1276957" [ 956.353045] env[61867]: _type = "Task" [ 956.353045] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.364030] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276957, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.427446] env[61867]: DEBUG nova.network.neutron [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Updating instance_info_cache with network_info: [{"id": "e070a089-e0e7-4b11-9688-26d528b37e6a", "address": "fa:16:3e:7b:ac:a0", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape070a089-e0", "ovs_interfaceid": "e070a089-e0e7-4b11-9688-26d528b37e6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.627694] env[61867]: DEBUG oslo_vmware.api [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276954, 'name': PowerOnVM_Task, 'duration_secs': 1.336283} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.627974] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 956.628039] env[61867]: INFO nova.compute.manager [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Took 8.57 seconds to spawn the instance on the hypervisor. [ 956.628226] env[61867]: DEBUG nova.compute.manager [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 956.629016] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9a5d12-8284-4690-896b-9dae225aeddc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.750778] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2dc6105d-d3ba-4efa-a228-74988815f9bc tempest-AttachVolumeShelveTestJSON-670241178 tempest-AttachVolumeShelveTestJSON-670241178-project-member] Lock "1df8427c-e75d-4b60-a92a-b5ba76b67081" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.052s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.865844] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276957, 'name': Rename_Task, 'duration_secs': 0.18662} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.866120] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 956.866418] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-efb5bdbe-5872-4181-829e-c105d947b49d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.877017] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 956.877017] env[61867]: value = "task-1276958" [ 956.877017] env[61867]: _type = "Task" [ 956.877017] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.896251] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276958, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.930193] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "refresh_cache-6e2136e7-c7e0-4a98-9899-f79c10f0e703" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.930645] env[61867]: DEBUG nova.compute.manager [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Instance network_info: |[{"id": "e070a089-e0e7-4b11-9688-26d528b37e6a", "address": "fa:16:3e:7b:ac:a0", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape070a089-e0", "ovs_interfaceid": "e070a089-e0e7-4b11-9688-26d528b37e6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 956.931513] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:ac:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f2020f39-42c4-4481-85c5-aaf03854b459', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e070a089-e0e7-4b11-9688-26d528b37e6a', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 956.939209] env[61867]: DEBUG oslo.service.loopingcall [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.939305] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 956.940518] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-005fe8ab-ca88-421b-bf18-fc391b63d0ef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.956156] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9cb32f0-eb60-437c-b632-4c4c101c788c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.964626] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48591835-4de8-4f4f-bcbe-af3826f2e74e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.968963] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 956.968963] env[61867]: value = "task-1276959" [ 956.968963] env[61867]: _type = "Task" [ 956.968963] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.002861] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3448e5c-8a89-446f-8ee8-6503c7b5ed9a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.005580] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276959, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.012365] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31bce299-a50d-4abe-8d63-6610d1241c70 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.028130] env[61867]: DEBUG nova.compute.provider_tree [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.146171] env[61867]: DEBUG nova.compute.manager [req-e8e72d4d-1ef9-45a2-8556-a834597eade9 req-f3801049-ac60-4bdb-9637-7b83fd61ee8a service nova] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Received event network-changed-e070a089-e0e7-4b11-9688-26d528b37e6a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 957.146400] env[61867]: DEBUG nova.compute.manager [req-e8e72d4d-1ef9-45a2-8556-a834597eade9 req-f3801049-ac60-4bdb-9637-7b83fd61ee8a service nova] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Refreshing instance network info cache due to event network-changed-e070a089-e0e7-4b11-9688-26d528b37e6a. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 957.146624] env[61867]: DEBUG oslo_concurrency.lockutils [req-e8e72d4d-1ef9-45a2-8556-a834597eade9 req-f3801049-ac60-4bdb-9637-7b83fd61ee8a service nova] Acquiring lock "refresh_cache-6e2136e7-c7e0-4a98-9899-f79c10f0e703" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.147207] env[61867]: DEBUG oslo_concurrency.lockutils [req-e8e72d4d-1ef9-45a2-8556-a834597eade9 req-f3801049-ac60-4bdb-9637-7b83fd61ee8a service nova] Acquired lock "refresh_cache-6e2136e7-c7e0-4a98-9899-f79c10f0e703" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.147207] env[61867]: DEBUG nova.network.neutron [req-e8e72d4d-1ef9-45a2-8556-a834597eade9 req-f3801049-ac60-4bdb-9637-7b83fd61ee8a service nova] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Refreshing network info cache for port e070a089-e0e7-4b11-9688-26d528b37e6a {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 957.150858] env[61867]: INFO nova.compute.manager [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Took 16.88 seconds to build instance. [ 957.388538] env[61867]: DEBUG oslo_vmware.api [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276958, 'name': PowerOnVM_Task, 'duration_secs': 0.51063} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.388538] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 957.388844] env[61867]: INFO nova.compute.manager [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Took 6.95 seconds to spawn the instance on the hypervisor. [ 957.388993] env[61867]: DEBUG nova.compute.manager [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 957.390141] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ae3bc7-1c16-42c8-b400-7bcc17aabd21 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.480254] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276959, 'name': CreateVM_Task} progress is 25%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.532071] env[61867]: DEBUG nova.scheduler.client.report [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 957.655126] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9652f11c-7624-4a8d-a62f-624ee9ae2b12 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "43dc6520-5256-4b5f-a273-6c9e0e407c72" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.394s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.914032] env[61867]: INFO nova.compute.manager [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Took 17.63 seconds to build instance. [ 957.939776] env[61867]: DEBUG nova.network.neutron [req-e8e72d4d-1ef9-45a2-8556-a834597eade9 req-f3801049-ac60-4bdb-9637-7b83fd61ee8a service nova] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Updated VIF entry in instance network info cache for port e070a089-e0e7-4b11-9688-26d528b37e6a. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 957.940180] env[61867]: DEBUG nova.network.neutron [req-e8e72d4d-1ef9-45a2-8556-a834597eade9 req-f3801049-ac60-4bdb-9637-7b83fd61ee8a service nova] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Updating instance_info_cache with network_info: [{"id": "e070a089-e0e7-4b11-9688-26d528b37e6a", "address": "fa:16:3e:7b:ac:a0", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape070a089-e0", "ovs_interfaceid": "e070a089-e0e7-4b11-9688-26d528b37e6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.980283] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276959, 'name': CreateVM_Task, 'duration_secs': 0.723092} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.980353] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 957.981074] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.981207] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.981532] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 957.981792] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31032787-2f26-4103-a578-3326e51238b3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.986871] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 957.986871] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e07f50-064c-5479-cb7c-81d396e74a79" [ 957.986871] env[61867]: _type = "Task" [ 957.986871] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.994969] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e07f50-064c-5479-cb7c-81d396e74a79, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.036503] env[61867]: DEBUG oslo_concurrency.lockutils [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.821s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.038789] env[61867]: DEBUG oslo_concurrency.lockutils [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 7.111s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.058852] env[61867]: INFO nova.scheduler.client.report [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Deleted allocations for instance 816ca796-d8ed-4843-9b1f-f169f48ff047 [ 958.415854] env[61867]: DEBUG oslo_concurrency.lockutils [None req-94be4bab-7be2-44a3-bf47-7d14e78580d6 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "5cf15593-54f6-405c-8435-2e3b378983b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.146s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.443688] env[61867]: DEBUG oslo_concurrency.lockutils [req-e8e72d4d-1ef9-45a2-8556-a834597eade9 req-f3801049-ac60-4bdb-9637-7b83fd61ee8a service nova] Releasing lock "refresh_cache-6e2136e7-c7e0-4a98-9899-f79c10f0e703" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.507904] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e07f50-064c-5479-cb7c-81d396e74a79, 'name': SearchDatastore_Task, 'duration_secs': 0.039162} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.508286] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.508537] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 958.508780] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.508966] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.509223] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 958.509562] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-910a369b-4260-4990-90a4-ed07dc1a7373 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.521484] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 958.521683] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 958.522805] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7c8ee20-f0e1-4171-94c8-cf40ffc6e670 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.529303] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 958.529303] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522cb0df-7b33-7b36-4ab0-d71e3d94a289" [ 958.529303] env[61867]: _type = "Task" [ 958.529303] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.541054] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522cb0df-7b33-7b36-4ab0-d71e3d94a289, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.576400] env[61867]: DEBUG oslo_concurrency.lockutils [None req-32bbca87-422e-4671-9408-93798d0c85b3 tempest-FloatingIPsAssociationTestJSON-1703938388 tempest-FloatingIPsAssociationTestJSON-1703938388-project-member] Lock "816ca796-d8ed-4843-9b1f-f169f48ff047" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.230s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.667274] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "43dc6520-5256-4b5f-a273-6c9e0e407c72" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.667620] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "43dc6520-5256-4b5f-a273-6c9e0e407c72" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.668323] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "43dc6520-5256-4b5f-a273-6c9e0e407c72-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.668323] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "43dc6520-5256-4b5f-a273-6c9e0e407c72-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.668323] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "43dc6520-5256-4b5f-a273-6c9e0e407c72-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.671129] env[61867]: INFO nova.compute.manager [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Terminating instance [ 958.673178] env[61867]: DEBUG nova.compute.manager [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 958.673388] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 958.677060] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0529ead8-fece-47a3-9e0f-15b96da11392 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.686504] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 958.686888] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f2a62edd-a084-4ef5-a772-a7cedaf5f4ef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.694896] env[61867]: DEBUG oslo_vmware.api [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 958.694896] env[61867]: value = "task-1276960" [ 958.694896] env[61867]: _type = "Task" [ 958.694896] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.717552] env[61867]: DEBUG oslo_vmware.api [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276960, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.801209] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e02b27-b6aa-43b2-b676-68121eaa455a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.810269] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d42bd1cd-999a-4813-b079-2e9aa7dcb7eb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.841168] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb05b63-62a2-495f-8f8f-9decace09ec7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.849141] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4854f0-2ce8-41c3-ab90-93d18e3d17e1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.864175] env[61867]: DEBUG nova.compute.provider_tree [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.023340] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d9819efd-a78e-44fa-953b-f9c151b3b0ab tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "5cf15593-54f6-405c-8435-2e3b378983b7" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.023635] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d9819efd-a78e-44fa-953b-f9c151b3b0ab tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "5cf15593-54f6-405c-8435-2e3b378983b7" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.023830] env[61867]: DEBUG nova.compute.manager [None req-d9819efd-a78e-44fa-953b-f9c151b3b0ab tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 959.024743] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef28e3b-8930-4a49-8962-6ed375b56e7e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.034729] env[61867]: DEBUG nova.compute.manager [None req-d9819efd-a78e-44fa-953b-f9c151b3b0ab tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61867) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 959.035315] env[61867]: DEBUG nova.objects.instance [None req-d9819efd-a78e-44fa-953b-f9c151b3b0ab tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lazy-loading 'flavor' on Instance uuid 5cf15593-54f6-405c-8435-2e3b378983b7 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.042326] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522cb0df-7b33-7b36-4ab0-d71e3d94a289, 'name': SearchDatastore_Task, 'duration_secs': 0.035742} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.043079] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-502fa252-1251-4694-9892-c8dbb5f4f32c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.049629] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 959.049629] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b47cc0-20b0-bd60-3d18-3eec9abcadec" [ 959.049629] env[61867]: _type = "Task" [ 959.049629] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.058013] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b47cc0-20b0-bd60-3d18-3eec9abcadec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.152733] env[61867]: DEBUG oslo_concurrency.lockutils [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "6d2dab88-4165-4952-8019-2eaf3b863115" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.153008] env[61867]: DEBUG oslo_concurrency.lockutils [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "6d2dab88-4165-4952-8019-2eaf3b863115" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.153279] env[61867]: DEBUG oslo_concurrency.lockutils [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "6d2dab88-4165-4952-8019-2eaf3b863115-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.153478] env[61867]: DEBUG oslo_concurrency.lockutils [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "6d2dab88-4165-4952-8019-2eaf3b863115-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.153674] env[61867]: DEBUG oslo_concurrency.lockutils [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "6d2dab88-4165-4952-8019-2eaf3b863115-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.155845] env[61867]: INFO nova.compute.manager [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Terminating instance [ 959.157546] env[61867]: DEBUG nova.compute.manager [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 959.157736] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 959.158564] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca2b6167-9999-4ecc-8609-78b836b58795 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.169958] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 959.169958] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-855a7a65-7136-42b5-b21e-4a4751ce885c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.176454] env[61867]: DEBUG oslo_vmware.api [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 959.176454] env[61867]: value = "task-1276961" [ 959.176454] env[61867]: _type = "Task" [ 959.176454] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.184561] env[61867]: DEBUG oslo_vmware.api [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276961, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.207473] env[61867]: DEBUG oslo_vmware.api [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276960, 'name': PowerOffVM_Task, 'duration_secs': 0.333016} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.207690] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 959.207873] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 959.208128] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-896cb68e-7231-44b5-90ca-75a13e97342a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.288943] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 959.290709] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 959.290709] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleting the datastore file [datastore2] 43dc6520-5256-4b5f-a273-6c9e0e407c72 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 959.290709] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af428f0e-fe1a-4013-9a7f-264ecced7a36 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.301112] env[61867]: DEBUG oslo_vmware.api [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 959.301112] env[61867]: value = "task-1276963" [ 959.301112] env[61867]: _type = "Task" [ 959.301112] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.313129] env[61867]: DEBUG oslo_vmware.api [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276963, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.367019] env[61867]: DEBUG nova.scheduler.client.report [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 959.541707] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9819efd-a78e-44fa-953b-f9c151b3b0ab tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 959.542013] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f9b090c6-9ef5-47e7-ab14-14f1cadbeb0a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.549769] env[61867]: DEBUG oslo_vmware.api [None req-d9819efd-a78e-44fa-953b-f9c151b3b0ab tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 959.549769] env[61867]: value = "task-1276964" [ 959.549769] env[61867]: _type = "Task" [ 959.549769] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.562749] env[61867]: DEBUG oslo_vmware.api [None req-d9819efd-a78e-44fa-953b-f9c151b3b0ab tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276964, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.566442] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b47cc0-20b0-bd60-3d18-3eec9abcadec, 'name': SearchDatastore_Task, 'duration_secs': 0.021617} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.566712] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.566980] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 6e2136e7-c7e0-4a98-9899-f79c10f0e703/6e2136e7-c7e0-4a98-9899-f79c10f0e703.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 959.567258] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fbb2c54e-09fb-4188-944a-3d7245dd0036 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.575628] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 959.575628] env[61867]: value = "task-1276965" [ 959.575628] env[61867]: _type = "Task" [ 959.575628] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.585631] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276965, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.695952] env[61867]: DEBUG oslo_vmware.api [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276961, 'name': PowerOffVM_Task, 'duration_secs': 0.237399} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.695952] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 959.695952] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 959.695952] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-141347c2-eab6-4a65-b3df-9b25ce27114d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.808946] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 959.809365] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 959.809606] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Deleting the datastore file [datastore2] 6d2dab88-4165-4952-8019-2eaf3b863115 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 959.810573] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d57be17b-c475-47b7-8a61-499187a6f23d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.816652] env[61867]: DEBUG oslo_vmware.api [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1276963, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.18857} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.817462] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.818081] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 959.818081] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 959.820826] env[61867]: INFO nova.compute.manager [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Took 1.14 seconds to destroy the instance on the hypervisor. [ 959.820826] env[61867]: DEBUG oslo.service.loopingcall [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.820826] env[61867]: DEBUG nova.compute.manager [-] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 959.820826] env[61867]: DEBUG nova.network.neutron [-] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 959.823516] env[61867]: DEBUG oslo_vmware.api [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for the task: (returnval){ [ 959.823516] env[61867]: value = "task-1276967" [ 959.823516] env[61867]: _type = "Task" [ 959.823516] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.834358] env[61867]: DEBUG oslo_vmware.api [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276967, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.066240] env[61867]: DEBUG oslo_vmware.api [None req-d9819efd-a78e-44fa-953b-f9c151b3b0ab tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276964, 'name': PowerOffVM_Task, 'duration_secs': 0.200896} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.068474] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9819efd-a78e-44fa-953b-f9c151b3b0ab tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 960.068731] env[61867]: DEBUG nova.compute.manager [None req-d9819efd-a78e-44fa-953b-f9c151b3b0ab tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 960.069605] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246ef734-ae27-4a6b-a6ff-9f6967329f40 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.089326] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276965, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482461} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.089619] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 6e2136e7-c7e0-4a98-9899-f79c10f0e703/6e2136e7-c7e0-4a98-9899-f79c10f0e703.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 960.089961] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 960.090255] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-19119bf3-3adc-461d-9b6d-8f4a9757e735 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.099120] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 960.099120] env[61867]: value = "task-1276968" [ 960.099120] env[61867]: _type = "Task" [ 960.099120] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.110576] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276968, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.338976] env[61867]: DEBUG oslo_vmware.api [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Task: {'id': task-1276967, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.338976} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.339507] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 960.339925] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 960.340295] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 960.343026] env[61867]: INFO nova.compute.manager [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Took 1.18 seconds to destroy the instance on the hypervisor. [ 960.343026] env[61867]: DEBUG oslo.service.loopingcall [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 960.343026] env[61867]: DEBUG nova.compute.manager [-] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 960.343026] env[61867]: DEBUG nova.network.neutron [-] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 960.379485] env[61867]: DEBUG oslo_concurrency.lockutils [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.341s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.380128] env[61867]: DEBUG nova.compute.manager [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=61867) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 960.386163] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.659s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.386163] env[61867]: INFO nova.compute.claims [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 960.396592] env[61867]: DEBUG nova.compute.manager [req-462e94c0-3b8e-4dbb-a0d8-e84df31a7bbd req-1c45bd41-b11e-4c92-b3a0-8b0090907561 service nova] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Received event network-vif-deleted-df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.399015] env[61867]: INFO nova.compute.manager [req-462e94c0-3b8e-4dbb-a0d8-e84df31a7bbd req-1c45bd41-b11e-4c92-b3a0-8b0090907561 service nova] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Neutron deleted interface df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83; detaching it from the instance and deleting it from the info cache [ 960.399015] env[61867]: DEBUG nova.network.neutron [req-462e94c0-3b8e-4dbb-a0d8-e84df31a7bbd req-1c45bd41-b11e-4c92-b3a0-8b0090907561 service nova] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.588528] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d9819efd-a78e-44fa-953b-f9c151b3b0ab tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "5cf15593-54f6-405c-8435-2e3b378983b7" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.565s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.615200] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276968, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073296} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.616256] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 960.617223] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff76639-8063-49cc-a96a-b7dc4cecb283 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.646630] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 6e2136e7-c7e0-4a98-9899-f79c10f0e703/6e2136e7-c7e0-4a98-9899-f79c10f0e703.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 960.650019] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de6413c9-1ab7-42b6-b273-8ddff38849e4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.670899] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 960.670899] env[61867]: value = "task-1276969" [ 960.670899] env[61867]: _type = "Task" [ 960.670899] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.680572] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276969, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.862366] env[61867]: DEBUG nova.compute.manager [req-e423dae8-9dda-455b-9ce1-ccd505edd928 req-ab64e4d7-2df8-43f2-b505-d763ac96cdea service nova] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Received event network-vif-deleted-101c4e9b-0368-4c02-9297-4eb478b6fc2b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.862629] env[61867]: INFO nova.compute.manager [req-e423dae8-9dda-455b-9ce1-ccd505edd928 req-ab64e4d7-2df8-43f2-b505-d763ac96cdea service nova] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Neutron deleted interface 101c4e9b-0368-4c02-9297-4eb478b6fc2b; detaching it from the instance and deleting it from the info cache [ 960.862872] env[61867]: DEBUG nova.network.neutron [req-e423dae8-9dda-455b-9ce1-ccd505edd928 req-ab64e4d7-2df8-43f2-b505-d763ac96cdea service nova] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.871260] env[61867]: DEBUG nova.network.neutron [-] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.900800] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f3bce7ac-c7c7-4ac7-898d-aaea907d1a62 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.915947] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f14cb3a-b463-4572-8b1b-ef02b59cff84 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.668086] env[61867]: DEBUG nova.objects.instance [None req-271a4dbe-a2d6-44c9-a672-f71952bc1e65 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lazy-loading 'flavor' on Instance uuid abb41c0c-6d0d-4147-a4af-554ab7d9e921 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.669276] env[61867]: DEBUG nova.network.neutron [-] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.671311] env[61867]: INFO nova.compute.manager [-] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Took 1.85 seconds to deallocate network for instance. [ 961.674025] env[61867]: DEBUG nova.compute.manager [req-462e94c0-3b8e-4dbb-a0d8-e84df31a7bbd req-1c45bd41-b11e-4c92-b3a0-8b0090907561 service nova] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Detach interface failed, port_id=df8e3f7f-3b22-484e-94b4-2aa1a2e9bf83, reason: Instance 43dc6520-5256-4b5f-a273-6c9e0e407c72 could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 961.675404] env[61867]: INFO nova.scheduler.client.report [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Deleted allocation for migration ca3d0401-216e-4d3c-a441-9d1297b25042 [ 961.682201] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5a5a4c9b-5876-4319-9012-3b2447de6184 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.699197] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276969, 'name': ReconfigVM_Task, 'duration_secs': 0.427098} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.702940] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 6e2136e7-c7e0-4a98-9899-f79c10f0e703/6e2136e7-c7e0-4a98-9899-f79c10f0e703.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 961.705686] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-37b3214d-9a5a-474b-926d-aafc1ff56cbe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.710351] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-166fcad1-108a-4f5d-a686-1b048a5de40b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.729711] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 961.729711] env[61867]: value = "task-1276970" [ 961.729711] env[61867]: _type = "Task" [ 961.729711] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.742744] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276970, 'name': Rename_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.751320] env[61867]: DEBUG nova.compute.manager [req-e423dae8-9dda-455b-9ce1-ccd505edd928 req-ab64e4d7-2df8-43f2-b505-d763ac96cdea service nova] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Detach interface failed, port_id=101c4e9b-0368-4c02-9297-4eb478b6fc2b, reason: Instance 6d2dab88-4165-4952-8019-2eaf3b863115 could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 961.868502] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c64853-3195-4e6d-bc09-65eb0185fc0f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.876211] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5aae31-6a51-4f67-af6e-6899449a49c2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.908747] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6086b03c-aab6-4758-bcf7-8c547d81d434 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.916774] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c91327-d3ba-4281-86aa-1dc81956e3b3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.932032] env[61867]: DEBUG nova.compute.provider_tree [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.175898] env[61867]: INFO nova.compute.manager [-] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Took 1.83 seconds to deallocate network for instance. [ 962.182285] env[61867]: DEBUG oslo_concurrency.lockutils [None req-271a4dbe-a2d6-44c9-a672-f71952bc1e65 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.182477] env[61867]: DEBUG oslo_concurrency.lockutils [None req-271a4dbe-a2d6-44c9-a672-f71952bc1e65 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.182652] env[61867]: DEBUG nova.network.neutron [None req-271a4dbe-a2d6-44c9-a672-f71952bc1e65 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 962.182828] env[61867]: DEBUG nova.objects.instance [None req-271a4dbe-a2d6-44c9-a672-f71952bc1e65 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lazy-loading 'info_cache' on Instance uuid abb41c0c-6d0d-4147-a4af-554ab7d9e921 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.188522] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "5cf15593-54f6-405c-8435-2e3b378983b7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.189589] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "5cf15593-54f6-405c-8435-2e3b378983b7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.189589] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "5cf15593-54f6-405c-8435-2e3b378983b7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.189589] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "5cf15593-54f6-405c-8435-2e3b378983b7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.189589] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "5cf15593-54f6-405c-8435-2e3b378983b7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.191276] env[61867]: DEBUG oslo_concurrency.lockutils [None req-45954f4e-915f-49c8-828c-485b717f1aaa tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 14.116s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.192248] env[61867]: INFO nova.compute.manager [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Terminating instance [ 962.193942] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.194503] env[61867]: DEBUG nova.compute.manager [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 962.194701] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 962.195829] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab325977-6856-4210-81b7-92f52523c217 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.206085] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 962.206085] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9cc164de-774d-4b80-92dc-2daa51edf6a4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.240473] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276970, 'name': Rename_Task, 'duration_secs': 0.14082} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.240772] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 962.241033] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7b00567a-572e-4d54-82d6-c0eab2078cdd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.247918] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 962.247918] env[61867]: value = "task-1276972" [ 962.247918] env[61867]: _type = "Task" [ 962.247918] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.256615] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276972, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.275953] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 962.276223] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 962.276460] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleting the datastore file [datastore2] 5cf15593-54f6-405c-8435-2e3b378983b7 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 962.276771] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b8d5ce24-e6a9-498e-a693-2ad92bbfc4f5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.283806] env[61867]: DEBUG oslo_vmware.api [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 962.283806] env[61867]: value = "task-1276973" [ 962.283806] env[61867]: _type = "Task" [ 962.283806] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.296295] env[61867]: DEBUG oslo_vmware.api [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276973, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.436045] env[61867]: DEBUG nova.scheduler.client.report [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.688246] env[61867]: DEBUG oslo_concurrency.lockutils [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.688246] env[61867]: DEBUG nova.objects.base [None req-271a4dbe-a2d6-44c9-a672-f71952bc1e65 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61867) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 962.759036] env[61867]: DEBUG oslo_vmware.api [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276972, 'name': PowerOnVM_Task, 'duration_secs': 0.461126} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.759570] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 962.759746] env[61867]: INFO nova.compute.manager [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Took 8.18 seconds to spawn the instance on the hypervisor. [ 962.759847] env[61867]: DEBUG nova.compute.manager [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 962.760785] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7fedcc4-cfee-44ba-8ca9-0e73bb9fcba6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.793993] env[61867]: DEBUG oslo_vmware.api [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1276973, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154044} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.794249] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 962.794461] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 962.795072] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 962.795072] env[61867]: INFO nova.compute.manager [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Took 0.60 seconds to destroy the instance on the hypervisor. [ 962.795072] env[61867]: DEBUG oslo.service.loopingcall [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.795318] env[61867]: DEBUG nova.compute.manager [-] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 962.795373] env[61867]: DEBUG nova.network.neutron [-] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 962.941422] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.557s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.941956] env[61867]: DEBUG nova.compute.manager [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 962.947926] env[61867]: DEBUG oslo_concurrency.lockutils [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.700s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.947926] env[61867]: INFO nova.compute.claims [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 963.237188] env[61867]: DEBUG nova.compute.manager [req-8cc778d2-f9de-4988-bf0f-d9dce49d7c57 req-6741343e-2e8d-4771-8c92-2e6387cc852f service nova] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Received event network-vif-deleted-a64da4a0-c1d4-4c3a-8386-a39668fa32e9 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.237508] env[61867]: INFO nova.compute.manager [req-8cc778d2-f9de-4988-bf0f-d9dce49d7c57 req-6741343e-2e8d-4771-8c92-2e6387cc852f service nova] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Neutron deleted interface a64da4a0-c1d4-4c3a-8386-a39668fa32e9; detaching it from the instance and deleting it from the info cache [ 963.237607] env[61867]: DEBUG nova.network.neutron [req-8cc778d2-f9de-4988-bf0f-d9dce49d7c57 req-6741343e-2e8d-4771-8c92-2e6387cc852f service nova] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.278424] env[61867]: INFO nova.compute.manager [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Took 16.86 seconds to build instance. [ 963.450246] env[61867]: DEBUG nova.compute.utils [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 963.454651] env[61867]: DEBUG nova.compute.manager [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 963.454651] env[61867]: DEBUG nova.network.neutron [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 963.516866] env[61867]: DEBUG nova.network.neutron [None req-271a4dbe-a2d6-44c9-a672-f71952bc1e65 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updating instance_info_cache with network_info: [{"id": "f0eb4126-40e7-4fe8-b276-192b91388aba", "address": "fa:16:3e:69:84:03", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0eb4126-40", "ovs_interfaceid": "f0eb4126-40e7-4fe8-b276-192b91388aba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.699678] env[61867]: DEBUG nova.policy [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ea634246436422a9407f82692b69ede', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '86ced20eaf4740e298dc6f8ca5550c09', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 963.700939] env[61867]: DEBUG nova.network.neutron [-] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.742639] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2204a64f-9217-4d15-aa7c-326d455c9e86 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.753996] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63c40eb-6918-444b-8d81-4c4aa2665c16 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.786859] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d11d84eb-e41d-49bb-9905-573d87cfe26b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.384s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.787232] env[61867]: DEBUG nova.compute.manager [req-8cc778d2-f9de-4988-bf0f-d9dce49d7c57 req-6741343e-2e8d-4771-8c92-2e6387cc852f service nova] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Detach interface failed, port_id=a64da4a0-c1d4-4c3a-8386-a39668fa32e9, reason: Instance 5cf15593-54f6-405c-8435-2e3b378983b7 could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 963.955974] env[61867]: DEBUG nova.compute.manager [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 964.020489] env[61867]: DEBUG oslo_concurrency.lockutils [None req-271a4dbe-a2d6-44c9-a672-f71952bc1e65 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "refresh_cache-abb41c0c-6d0d-4147-a4af-554ab7d9e921" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.046864] env[61867]: DEBUG nova.network.neutron [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Successfully created port: a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 964.203309] env[61867]: INFO nova.compute.manager [-] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Took 1.41 seconds to deallocate network for instance. [ 964.207337] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f83184-790c-4af9-8f75-adc2affb7ca1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.225806] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24facd81-b343-4dca-9734-783f6fb9a9b5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.263213] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8d4ad1-53f3-4edc-97d6-58cde24740d1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.272653] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a0a0f73-8191-4284-b2a2-58ccd7913a7c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.289435] env[61867]: DEBUG nova.compute.provider_tree [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 964.523802] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-271a4dbe-a2d6-44c9-a672-f71952bc1e65 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 964.524337] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-97d149f1-c664-416e-825a-24e7a144d0bc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.532960] env[61867]: DEBUG oslo_vmware.api [None req-271a4dbe-a2d6-44c9-a672-f71952bc1e65 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 964.532960] env[61867]: value = "task-1276974" [ 964.532960] env[61867]: _type = "Task" [ 964.532960] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.545244] env[61867]: DEBUG oslo_vmware.api [None req-271a4dbe-a2d6-44c9-a672-f71952bc1e65 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276974, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.717435] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.813100] env[61867]: ERROR nova.scheduler.client.report [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [req-132d8c92-ef61-48a0-9b5f-354aec280036] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 25720271-a549-4916-abe3-e5ed9b765889. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-132d8c92-ef61-48a0-9b5f-354aec280036"}]} [ 964.840480] env[61867]: DEBUG nova.scheduler.client.report [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Refreshing inventories for resource provider 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 964.853829] env[61867]: DEBUG nova.scheduler.client.report [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Updating ProviderTree inventory for provider 25720271-a549-4916-abe3-e5ed9b765889 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 964.854082] env[61867]: DEBUG nova.compute.provider_tree [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 964.869671] env[61867]: DEBUG nova.scheduler.client.report [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Refreshing aggregate associations for resource provider 25720271-a549-4916-abe3-e5ed9b765889, aggregates: None {{(pid=61867) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 964.889015] env[61867]: DEBUG nova.scheduler.client.report [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Refreshing trait associations for resource provider 25720271-a549-4916-abe3-e5ed9b765889, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61867) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 964.901988] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Acquiring lock "d19e7706-e6ac-47dd-8ec3-c30bd315559d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.902337] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Lock "d19e7706-e6ac-47dd-8ec3-c30bd315559d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.970185] env[61867]: DEBUG nova.compute.manager [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 965.005710] env[61867]: DEBUG nova.virt.hardware [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 965.005965] env[61867]: DEBUG nova.virt.hardware [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 965.006147] env[61867]: DEBUG nova.virt.hardware [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 965.006338] env[61867]: DEBUG nova.virt.hardware [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 965.006490] env[61867]: DEBUG nova.virt.hardware [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 965.006641] env[61867]: DEBUG nova.virt.hardware [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 965.006851] env[61867]: DEBUG nova.virt.hardware [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 965.007341] env[61867]: DEBUG nova.virt.hardware [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 965.007586] env[61867]: DEBUG nova.virt.hardware [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 965.008077] env[61867]: DEBUG nova.virt.hardware [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 965.008077] env[61867]: DEBUG nova.virt.hardware [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 965.011355] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ed4a1e-0941-4877-a898-470531ef89e1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.020548] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eee4989-e7ac-4b8f-a242-6402a060e0ed {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.046115] env[61867]: DEBUG oslo_vmware.api [None req-271a4dbe-a2d6-44c9-a672-f71952bc1e65 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276974, 'name': PowerOnVM_Task, 'duration_secs': 0.419012} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.048649] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-271a4dbe-a2d6-44c9-a672-f71952bc1e65 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 965.048843] env[61867]: DEBUG nova.compute.manager [None req-271a4dbe-a2d6-44c9-a672-f71952bc1e65 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 965.050104] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83c02d3-4c76-460e-81a5-dc2061ba722f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.127798] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c7d5abf-9f03-4020-8894-914a3b0aeeb3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.139202] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87efae53-14fc-4775-8a6b-4eda36afc24f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.172231] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fda7d70-ceb9-41db-89a2-01d2fa08f3f6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.180932] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45ce91c5-56e2-4d23-ac79-fa74ff980be3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.196462] env[61867]: DEBUG nova.compute.provider_tree [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 965.405163] env[61867]: DEBUG nova.compute.manager [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 965.506838] env[61867]: DEBUG nova.compute.manager [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Stashing vm_state: active {{(pid=61867) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 965.613414] env[61867]: DEBUG nova.compute.manager [req-16513ea2-fbe0-4534-bd0b-8963676d770f req-720fe6ec-e1e3-4bc5-836f-16a47ee5751a service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Received event network-vif-plugged-a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.613414] env[61867]: DEBUG oslo_concurrency.lockutils [req-16513ea2-fbe0-4534-bd0b-8963676d770f req-720fe6ec-e1e3-4bc5-836f-16a47ee5751a service nova] Acquiring lock "0ae83268-ef50-42e6-954d-5e67c4afc20e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.613414] env[61867]: DEBUG oslo_concurrency.lockutils [req-16513ea2-fbe0-4534-bd0b-8963676d770f req-720fe6ec-e1e3-4bc5-836f-16a47ee5751a service nova] Lock "0ae83268-ef50-42e6-954d-5e67c4afc20e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.613654] env[61867]: DEBUG oslo_concurrency.lockutils [req-16513ea2-fbe0-4534-bd0b-8963676d770f req-720fe6ec-e1e3-4bc5-836f-16a47ee5751a service nova] Lock "0ae83268-ef50-42e6-954d-5e67c4afc20e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.613899] env[61867]: DEBUG nova.compute.manager [req-16513ea2-fbe0-4534-bd0b-8963676d770f req-720fe6ec-e1e3-4bc5-836f-16a47ee5751a service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] No waiting events found dispatching network-vif-plugged-a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 965.613938] env[61867]: WARNING nova.compute.manager [req-16513ea2-fbe0-4534-bd0b-8963676d770f req-720fe6ec-e1e3-4bc5-836f-16a47ee5751a service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Received unexpected event network-vif-plugged-a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab for instance with vm_state building and task_state spawning. [ 965.709302] env[61867]: DEBUG nova.network.neutron [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Successfully updated port: a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 965.745569] env[61867]: DEBUG nova.scheduler.client.report [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Updated inventory for provider 25720271-a549-4916-abe3-e5ed9b765889 with generation 119 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 965.745866] env[61867]: DEBUG nova.compute.provider_tree [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Updating resource provider 25720271-a549-4916-abe3-e5ed9b765889 generation from 119 to 120 during operation: update_inventory {{(pid=61867) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 965.746070] env[61867]: DEBUG nova.compute.provider_tree [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 965.928165] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.026490] env[61867]: DEBUG oslo_concurrency.lockutils [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.215097] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.215097] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.215097] env[61867]: DEBUG nova.network.neutron [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 966.251355] env[61867]: DEBUG oslo_concurrency.lockutils [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.306s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.252062] env[61867]: DEBUG nova.compute.manager [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 966.254877] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 9.963s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.296482] env[61867]: DEBUG oslo_concurrency.lockutils [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.296736] env[61867]: DEBUG oslo_concurrency.lockutils [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.297131] env[61867]: DEBUG oslo_concurrency.lockutils [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.297473] env[61867]: DEBUG oslo_concurrency.lockutils [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.297652] env[61867]: DEBUG oslo_concurrency.lockutils [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.299898] env[61867]: INFO nova.compute.manager [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Terminating instance [ 966.302259] env[61867]: DEBUG nova.compute.manager [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 966.302506] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 966.303681] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df16f258-01be-45b4-afe8-3245b514864c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.315208] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 966.316751] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6fac9013-54cb-493a-ba50-c405afeb02a6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.326201] env[61867]: DEBUG oslo_vmware.api [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 966.326201] env[61867]: value = "task-1276975" [ 966.326201] env[61867]: _type = "Task" [ 966.326201] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.335914] env[61867]: DEBUG oslo_vmware.api [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276975, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.749403] env[61867]: DEBUG nova.network.neutron [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 966.758083] env[61867]: DEBUG nova.compute.utils [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 966.759816] env[61867]: DEBUG nova.compute.manager [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 966.760059] env[61867]: DEBUG nova.network.neutron [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 966.826312] env[61867]: DEBUG nova.policy [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5a066048582d4f7d905e201884470509', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '60b35760e3e14245aea3600d36c838dc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 966.838054] env[61867]: DEBUG oslo_vmware.api [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276975, 'name': PowerOffVM_Task, 'duration_secs': 0.407094} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.838345] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 966.838511] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 966.838776] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8a2bf2dd-ff72-4a2a-9019-3fae62e5ce1f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.895734] env[61867]: DEBUG nova.network.neutron [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Updating instance_info_cache with network_info: [{"id": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "address": "fa:16:3e:a0:23:06", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3537cac-b5", "ovs_interfaceid": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.151071] env[61867]: DEBUG nova.network.neutron [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Successfully created port: 86895cfd-1390-4fcd-921e-e39e1ac94da0 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 967.195601] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 967.195905] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 967.196222] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Deleting the datastore file [datastore1] abb41c0c-6d0d-4147-a4af-554ab7d9e921 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 967.196555] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d7001920-e5da-43df-b709-5a57a5929eb7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.203796] env[61867]: DEBUG oslo_vmware.api [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 967.203796] env[61867]: value = "task-1276977" [ 967.203796] env[61867]: _type = "Task" [ 967.203796] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.213166] env[61867]: DEBUG oslo_vmware.api [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276977, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.272043] env[61867]: DEBUG nova.compute.manager [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 967.279802] env[61867]: INFO nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Updating resource usage from migration 44205de9-d7fc-454a-8217-47bec978b3d8 [ 967.305863] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 6e41989e-b8fa-4009-af1e-1ce859b329a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 967.306083] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 8a83f4f2-58eb-473e-9b1e-32ce633554f9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 967.306250] env[61867]: WARNING nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 6d2dab88-4165-4952-8019-2eaf3b863115 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 967.306401] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 2aa08603-d87f-4734-bdfe-fdd610d54e1f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 967.306555] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 7479bf91-5aef-4e75-a127-7e82ae15a003 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 967.306706] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 8a7f4314-0fd4-49f9-8eb6-12baa0977a53 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 967.306842] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance abb41c0c-6d0d-4147-a4af-554ab7d9e921 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 967.306985] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 4d7bfb31-d565-49e4-8c51-0122acebba2d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 967.307179] env[61867]: WARNING nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 5cf15593-54f6-405c-8435-2e3b378983b7 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 967.307354] env[61867]: WARNING nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 43dc6520-5256-4b5f-a273-6c9e0e407c72 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 967.307542] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 0ae83268-ef50-42e6-954d-5e67c4afc20e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 967.307703] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance f426dd3e-2806-4d70-8784-75c611fd03e4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 967.398321] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.398694] env[61867]: DEBUG nova.compute.manager [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Instance network_info: |[{"id": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "address": "fa:16:3e:a0:23:06", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3537cac-b5", "ovs_interfaceid": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 967.399225] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:23:06', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '11da2092-76f7-447e-babb-8fc14ad39a71', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 967.407133] env[61867]: DEBUG oslo.service.loopingcall [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.408017] env[61867]: DEBUG nova.network.neutron [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Successfully created port: bb767aa8-7bfe-4604-9152-059b0eaa3311 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 967.409987] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 967.410260] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-886094c4-deec-401e-9803-df26d971ceca {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.430346] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 967.430346] env[61867]: value = "task-1276978" [ 967.430346] env[61867]: _type = "Task" [ 967.430346] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.438514] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276978, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.714068] env[61867]: DEBUG oslo_vmware.api [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1276977, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158332} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.714371] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 967.714543] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 967.714722] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 967.714917] env[61867]: INFO nova.compute.manager [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Took 1.41 seconds to destroy the instance on the hypervisor. [ 967.715194] env[61867]: DEBUG oslo.service.loopingcall [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.715393] env[61867]: DEBUG nova.compute.manager [-] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 967.715483] env[61867]: DEBUG nova.network.neutron [-] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 967.788086] env[61867]: DEBUG nova.compute.manager [req-d815b7da-da2e-408a-905e-8a6ce82baa16 req-87161b05-57c2-4b9e-b0e3-06ff07fd30fc service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Received event network-changed-a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.788309] env[61867]: DEBUG nova.compute.manager [req-d815b7da-da2e-408a-905e-8a6ce82baa16 req-87161b05-57c2-4b9e-b0e3-06ff07fd30fc service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Refreshing instance network info cache due to event network-changed-a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 967.788603] env[61867]: DEBUG oslo_concurrency.lockutils [req-d815b7da-da2e-408a-905e-8a6ce82baa16 req-87161b05-57c2-4b9e-b0e3-06ff07fd30fc service nova] Acquiring lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.788686] env[61867]: DEBUG oslo_concurrency.lockutils [req-d815b7da-da2e-408a-905e-8a6ce82baa16 req-87161b05-57c2-4b9e-b0e3-06ff07fd30fc service nova] Acquired lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.788822] env[61867]: DEBUG nova.network.neutron [req-d815b7da-da2e-408a-905e-8a6ce82baa16 req-87161b05-57c2-4b9e-b0e3-06ff07fd30fc service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Refreshing network info cache for port a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 967.811652] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance d19e7706-e6ac-47dd-8ec3-c30bd315559d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 967.811824] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Migration 44205de9-d7fc-454a-8217-47bec978b3d8 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 967.811952] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 6e2136e7-c7e0-4a98-9899-f79c10f0e703 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 967.812285] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 967.812449] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 967.949343] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276978, 'name': CreateVM_Task, 'duration_secs': 0.299202} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.949506] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 967.950700] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.950700] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.950976] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 967.951226] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4f251a9-1dc2-4fcf-b9bb-aaca772e630e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.959910] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 967.959910] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ea6bdd-6459-228e-28d5-e28534ed7c15" [ 967.959910] env[61867]: _type = "Task" [ 967.959910] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.973482] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ea6bdd-6459-228e-28d5-e28534ed7c15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.041429] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad70b69d-9c86-44cc-9726-866341378e75 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.049089] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc16edd7-dd75-42a7-a015-34aa21589c6b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.081382] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e61a6da-54b9-48ae-8fc6-ddfc28d2689a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.090274] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f469bab-7dd9-4f17-8912-9750cdeb0117 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.104327] env[61867]: DEBUG nova.compute.provider_tree [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.302793] env[61867]: DEBUG nova.compute.manager [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 968.329838] env[61867]: DEBUG nova.virt.hardware [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 968.330134] env[61867]: DEBUG nova.virt.hardware [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 968.330307] env[61867]: DEBUG nova.virt.hardware [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 968.330493] env[61867]: DEBUG nova.virt.hardware [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 968.330643] env[61867]: DEBUG nova.virt.hardware [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 968.330794] env[61867]: DEBUG nova.virt.hardware [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 968.331034] env[61867]: DEBUG nova.virt.hardware [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 968.331177] env[61867]: DEBUG nova.virt.hardware [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 968.331350] env[61867]: DEBUG nova.virt.hardware [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 968.331515] env[61867]: DEBUG nova.virt.hardware [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 968.331686] env[61867]: DEBUG nova.virt.hardware [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 968.333143] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1db80d1-d355-487b-befa-ad3899e7f2a2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.341393] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c44628-1678-415b-a591-9cd65cffa72c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.472280] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ea6bdd-6459-228e-28d5-e28534ed7c15, 'name': SearchDatastore_Task, 'duration_secs': 0.010455} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.472617] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.472851] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 968.473092] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.473244] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.473432] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 968.473692] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d9006947-ba37-4189-86a7-abf4014dd2e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.482747] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 968.482927] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 968.483658] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a54a1969-7b7b-431c-9b2d-8a0c3e9f66c2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.488690] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 968.488690] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e6950f-be81-8de3-f8d2-4f74b8669b76" [ 968.488690] env[61867]: _type = "Task" [ 968.488690] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.496477] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e6950f-be81-8de3-f8d2-4f74b8669b76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.497862] env[61867]: DEBUG nova.network.neutron [req-d815b7da-da2e-408a-905e-8a6ce82baa16 req-87161b05-57c2-4b9e-b0e3-06ff07fd30fc service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Updated VIF entry in instance network info cache for port a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 968.498203] env[61867]: DEBUG nova.network.neutron [req-d815b7da-da2e-408a-905e-8a6ce82baa16 req-87161b05-57c2-4b9e-b0e3-06ff07fd30fc service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Updating instance_info_cache with network_info: [{"id": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "address": "fa:16:3e:a0:23:06", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3537cac-b5", "ovs_interfaceid": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.607838] env[61867]: DEBUG nova.scheduler.client.report [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 968.680081] env[61867]: DEBUG nova.network.neutron [-] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.768675] env[61867]: DEBUG oslo_concurrency.lockutils [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "8a7f4314-0fd4-49f9-8eb6-12baa0977a53" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.769138] env[61867]: DEBUG oslo_concurrency.lockutils [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "8a7f4314-0fd4-49f9-8eb6-12baa0977a53" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.769375] env[61867]: DEBUG oslo_concurrency.lockutils [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "8a7f4314-0fd4-49f9-8eb6-12baa0977a53-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.769575] env[61867]: DEBUG oslo_concurrency.lockutils [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "8a7f4314-0fd4-49f9-8eb6-12baa0977a53-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.769751] env[61867]: DEBUG oslo_concurrency.lockutils [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "8a7f4314-0fd4-49f9-8eb6-12baa0977a53-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.772253] env[61867]: INFO nova.compute.manager [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Terminating instance [ 968.774053] env[61867]: DEBUG nova.compute.manager [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 968.774285] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 968.775140] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdd4ee12-a28c-4468-a090-ca22ca59e9e9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.783829] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 968.784079] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a9f4f68-101d-4fd1-904e-2fbc20be90eb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.791914] env[61867]: DEBUG oslo_vmware.api [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 968.791914] env[61867]: value = "task-1276979" [ 968.791914] env[61867]: _type = "Task" [ 968.791914] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.801210] env[61867]: DEBUG oslo_vmware.api [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276979, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.003384] env[61867]: DEBUG oslo_concurrency.lockutils [req-d815b7da-da2e-408a-905e-8a6ce82baa16 req-87161b05-57c2-4b9e-b0e3-06ff07fd30fc service nova] Releasing lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.003832] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e6950f-be81-8de3-f8d2-4f74b8669b76, 'name': SearchDatastore_Task, 'duration_secs': 0.011482} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.004672] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccf798b4-d47c-4bc8-8ef0-d9f337e2a4d3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.010818] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 969.010818] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e814a1-71ed-0dd1-7706-aa820fec6ce2" [ 969.010818] env[61867]: _type = "Task" [ 969.010818] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.019605] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e814a1-71ed-0dd1-7706-aa820fec6ce2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.113632] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61867) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 969.113866] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.859s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.114202] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.920s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.114390] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.117347] env[61867]: DEBUG oslo_concurrency.lockutils [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.431s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.117347] env[61867]: DEBUG oslo_concurrency.lockutils [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.120264] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.403s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.120456] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.122437] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.194s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.124635] env[61867]: INFO nova.compute.claims [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 969.134879] env[61867]: DEBUG nova.compute.manager [req-dd0cea33-8881-40f6-aad2-ddae10bd37df req-d7f3fe2c-76db-462b-8c18-d9f1322745d1 service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Received event network-vif-plugged-86895cfd-1390-4fcd-921e-e39e1ac94da0 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.135362] env[61867]: DEBUG oslo_concurrency.lockutils [req-dd0cea33-8881-40f6-aad2-ddae10bd37df req-d7f3fe2c-76db-462b-8c18-d9f1322745d1 service nova] Acquiring lock "f426dd3e-2806-4d70-8784-75c611fd03e4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.135610] env[61867]: DEBUG oslo_concurrency.lockutils [req-dd0cea33-8881-40f6-aad2-ddae10bd37df req-d7f3fe2c-76db-462b-8c18-d9f1322745d1 service nova] Lock "f426dd3e-2806-4d70-8784-75c611fd03e4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.135803] env[61867]: DEBUG oslo_concurrency.lockutils [req-dd0cea33-8881-40f6-aad2-ddae10bd37df req-d7f3fe2c-76db-462b-8c18-d9f1322745d1 service nova] Lock "f426dd3e-2806-4d70-8784-75c611fd03e4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.135982] env[61867]: DEBUG nova.compute.manager [req-dd0cea33-8881-40f6-aad2-ddae10bd37df req-d7f3fe2c-76db-462b-8c18-d9f1322745d1 service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] No waiting events found dispatching network-vif-plugged-86895cfd-1390-4fcd-921e-e39e1ac94da0 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 969.136180] env[61867]: WARNING nova.compute.manager [req-dd0cea33-8881-40f6-aad2-ddae10bd37df req-d7f3fe2c-76db-462b-8c18-d9f1322745d1 service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Received unexpected event network-vif-plugged-86895cfd-1390-4fcd-921e-e39e1ac94da0 for instance with vm_state building and task_state spawning. [ 969.151866] env[61867]: INFO nova.scheduler.client.report [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Deleted allocations for instance 6d2dab88-4165-4952-8019-2eaf3b863115 [ 969.161071] env[61867]: INFO nova.scheduler.client.report [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleted allocations for instance 43dc6520-5256-4b5f-a273-6c9e0e407c72 [ 969.172823] env[61867]: INFO nova.scheduler.client.report [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleted allocations for instance 5cf15593-54f6-405c-8435-2e3b378983b7 [ 969.182500] env[61867]: INFO nova.compute.manager [-] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Took 1.47 seconds to deallocate network for instance. [ 969.273261] env[61867]: DEBUG nova.network.neutron [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Successfully updated port: 86895cfd-1390-4fcd-921e-e39e1ac94da0 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 969.301865] env[61867]: DEBUG oslo_vmware.api [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276979, 'name': PowerOffVM_Task, 'duration_secs': 0.181696} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.302212] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 969.302438] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 969.302651] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-99e78267-ad4c-4b3e-8401-6a133c66fbb1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.367384] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 969.367724] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 969.367897] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Deleting the datastore file [datastore2] 8a7f4314-0fd4-49f9-8eb6-12baa0977a53 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 969.368181] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-18f4a696-e203-4867-a352-0ab3dfbfe66e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.374519] env[61867]: DEBUG oslo_vmware.api [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 969.374519] env[61867]: value = "task-1276981" [ 969.374519] env[61867]: _type = "Task" [ 969.374519] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.383111] env[61867]: DEBUG oslo_vmware.api [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276981, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.521296] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e814a1-71ed-0dd1-7706-aa820fec6ce2, 'name': SearchDatastore_Task, 'duration_secs': 0.015442} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.521620] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.521884] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 0ae83268-ef50-42e6-954d-5e67c4afc20e/0ae83268-ef50-42e6-954d-5e67c4afc20e.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 969.522197] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b68e0938-22cf-4d96-aab5-f3767171d6ea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.528929] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 969.528929] env[61867]: value = "task-1276982" [ 969.528929] env[61867]: _type = "Task" [ 969.528929] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.537208] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276982, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.667617] env[61867]: DEBUG oslo_concurrency.lockutils [None req-93593117-8fb1-41cd-bae3-725533f3ac99 tempest-VolumesAdminNegativeTest-1818537246 tempest-VolumesAdminNegativeTest-1818537246-project-member] Lock "6d2dab88-4165-4952-8019-2eaf3b863115" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.514s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.669760] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9bfe36d5-9884-4b4f-998a-4e57fe951d10 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "43dc6520-5256-4b5f-a273-6c9e0e407c72" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.682791] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c8bc5c6-6716-482d-b84c-c33626a8b559 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "5cf15593-54f6-405c-8435-2e3b378983b7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.493s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.692727] env[61867]: DEBUG oslo_concurrency.lockutils [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.827821] env[61867]: DEBUG nova.compute.manager [req-b46635ae-df3f-45b5-968c-9d056a171b3a req-39989e60-f24f-4acd-968e-a7fc29cc80c8 service nova] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Received event network-vif-deleted-f0eb4126-40e7-4fe8-b276-192b91388aba {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.886644] env[61867]: DEBUG oslo_vmware.api [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1276981, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132631} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.887459] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.887994] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 969.888326] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 969.888654] env[61867]: INFO nova.compute.manager [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Took 1.11 seconds to destroy the instance on the hypervisor. [ 969.889026] env[61867]: DEBUG oslo.service.loopingcall [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.889349] env[61867]: DEBUG nova.compute.manager [-] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 969.889567] env[61867]: DEBUG nova.network.neutron [-] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 970.044916] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276982, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.449505} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.044916] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 0ae83268-ef50-42e6-954d-5e67c4afc20e/0ae83268-ef50-42e6-954d-5e67c4afc20e.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 970.044916] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 970.044916] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-69f92ffa-9dfb-4396-a109-e14fc273447c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.053030] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 970.053030] env[61867]: value = "task-1276983" [ 970.053030] env[61867]: _type = "Task" [ 970.053030] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.061461] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276983, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.358856] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec35cd4-1edf-4436-a346-db6090dfedd3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.369169] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14ae5db-e247-411f-ad58-233785287017 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.411467] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bb819a2-d01f-4c49-aebf-13924a610c34 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.420829] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ba5b16-fb84-48dc-a797-6c501f6fb0c3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.438380] env[61867]: DEBUG nova.compute.provider_tree [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 970.562094] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276983, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073873} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.562531] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 970.563426] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1290dc7-ed86-4197-bed0-1a0702c1af78 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.587690] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 0ae83268-ef50-42e6-954d-5e67c4afc20e/0ae83268-ef50-42e6-954d-5e67c4afc20e.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 970.588337] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-800b15e9-7741-4ea6-89c3-ce15a34fd0e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.613107] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 970.613107] env[61867]: value = "task-1276984" [ 970.613107] env[61867]: _type = "Task" [ 970.613107] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.624941] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276984, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.882443] env[61867]: DEBUG oslo_concurrency.lockutils [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "976299f8-c6cb-4106-90f2-0a4c85625d8c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.882780] env[61867]: DEBUG oslo_concurrency.lockutils [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "976299f8-c6cb-4106-90f2-0a4c85625d8c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.959920] env[61867]: ERROR nova.scheduler.client.report [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [req-bc153b9b-0ce8-4afe-a10f-a22eb036e080] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 25720271-a549-4916-abe3-e5ed9b765889. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-bc153b9b-0ce8-4afe-a10f-a22eb036e080"}]} [ 970.979201] env[61867]: DEBUG nova.scheduler.client.report [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Refreshing inventories for resource provider 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 970.993974] env[61867]: DEBUG nova.scheduler.client.report [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Updating ProviderTree inventory for provider 25720271-a549-4916-abe3-e5ed9b765889 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 970.994642] env[61867]: DEBUG nova.compute.provider_tree [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 971.005302] env[61867]: DEBUG nova.scheduler.client.report [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Refreshing aggregate associations for resource provider 25720271-a549-4916-abe3-e5ed9b765889, aggregates: None {{(pid=61867) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 971.008343] env[61867]: DEBUG nova.network.neutron [-] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.036224] env[61867]: DEBUG nova.scheduler.client.report [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Refreshing trait associations for resource provider 25720271-a549-4916-abe3-e5ed9b765889, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61867) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 971.126195] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276984, 'name': ReconfigVM_Task, 'duration_secs': 0.305744} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.126195] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 0ae83268-ef50-42e6-954d-5e67c4afc20e/0ae83268-ef50-42e6-954d-5e67c4afc20e.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 971.126626] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-15efdfff-4f92-4711-9210-af53db718cd0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.135333] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 971.135333] env[61867]: value = "task-1276985" [ 971.135333] env[61867]: _type = "Task" [ 971.135333] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.144206] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276985, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.279553] env[61867]: DEBUG nova.compute.manager [req-935e0c3f-a3db-4813-937b-4524a5de6226 req-a6514bc1-22f5-45ef-abc6-1284f91b3d4a service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Received event network-changed-86895cfd-1390-4fcd-921e-e39e1ac94da0 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.279817] env[61867]: DEBUG nova.compute.manager [req-935e0c3f-a3db-4813-937b-4524a5de6226 req-a6514bc1-22f5-45ef-abc6-1284f91b3d4a service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Refreshing instance network info cache due to event network-changed-86895cfd-1390-4fcd-921e-e39e1ac94da0. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 971.279981] env[61867]: DEBUG oslo_concurrency.lockutils [req-935e0c3f-a3db-4813-937b-4524a5de6226 req-a6514bc1-22f5-45ef-abc6-1284f91b3d4a service nova] Acquiring lock "refresh_cache-f426dd3e-2806-4d70-8784-75c611fd03e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.280220] env[61867]: DEBUG oslo_concurrency.lockutils [req-935e0c3f-a3db-4813-937b-4524a5de6226 req-a6514bc1-22f5-45ef-abc6-1284f91b3d4a service nova] Acquired lock "refresh_cache-f426dd3e-2806-4d70-8784-75c611fd03e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.280338] env[61867]: DEBUG nova.network.neutron [req-935e0c3f-a3db-4813-937b-4524a5de6226 req-a6514bc1-22f5-45ef-abc6-1284f91b3d4a service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Refreshing network info cache for port 86895cfd-1390-4fcd-921e-e39e1ac94da0 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 971.288961] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5863cf8-be08-45eb-81ad-d96401788a66 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.297887] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92ac50c9-a060-410c-b0ed-fbd47cb67f08 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.330195] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc719748-4a1b-4b07-b4e0-73a917e3c782 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.338493] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f82ffdcd-7128-4399-96fe-5c3a3dc3ebbb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.356700] env[61867]: DEBUG nova.compute.provider_tree [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 971.386511] env[61867]: DEBUG nova.compute.manager [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 971.394501] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "4006cabd-cb6e-45b6-b87e-9b4b1550f4fe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.394755] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "4006cabd-cb6e-45b6-b87e-9b4b1550f4fe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.510444] env[61867]: INFO nova.compute.manager [-] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Took 1.62 seconds to deallocate network for instance. [ 971.651024] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276985, 'name': Rename_Task, 'duration_secs': 0.15527} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.651024] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 971.651024] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-38c8c2fe-6d4e-4f62-ada0-e43ca9ad46cf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.657145] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 971.657145] env[61867]: value = "task-1276986" [ 971.657145] env[61867]: _type = "Task" [ 971.657145] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.666890] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276986, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.817586] env[61867]: DEBUG nova.network.neutron [req-935e0c3f-a3db-4813-937b-4524a5de6226 req-a6514bc1-22f5-45ef-abc6-1284f91b3d4a service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 971.855938] env[61867]: DEBUG nova.compute.manager [req-2d9055cb-259b-46c7-853f-35d8bd4c19ce req-463f1031-b556-4406-97f3-a1a60e3d0da2 service nova] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Received event network-vif-deleted-211a1ec3-d618-4c1d-ac61-7e7b5681b9ed {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.890332] env[61867]: DEBUG nova.scheduler.client.report [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Updated inventory for provider 25720271-a549-4916-abe3-e5ed9b765889 with generation 122 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 971.890612] env[61867]: DEBUG nova.compute.provider_tree [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Updating resource provider 25720271-a549-4916-abe3-e5ed9b765889 generation from 122 to 123 during operation: update_inventory {{(pid=61867) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 971.890805] env[61867]: DEBUG nova.compute.provider_tree [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 971.899745] env[61867]: DEBUG nova.compute.manager [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 971.919702] env[61867]: DEBUG nova.network.neutron [req-935e0c3f-a3db-4813-937b-4524a5de6226 req-a6514bc1-22f5-45ef-abc6-1284f91b3d4a service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.926830] env[61867]: DEBUG oslo_concurrency.lockutils [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.990940] env[61867]: DEBUG nova.network.neutron [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Successfully updated port: bb767aa8-7bfe-4604-9152-059b0eaa3311 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 972.018021] env[61867]: DEBUG oslo_concurrency.lockutils [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.168175] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276986, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.200936] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 972.401578] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.279s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.402319] env[61867]: DEBUG nova.compute.manager [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 972.405640] env[61867]: DEBUG oslo_concurrency.lockutils [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.379s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.422618] env[61867]: DEBUG oslo_concurrency.lockutils [req-935e0c3f-a3db-4813-937b-4524a5de6226 req-a6514bc1-22f5-45ef-abc6-1284f91b3d4a service nova] Releasing lock "refresh_cache-f426dd3e-2806-4d70-8784-75c611fd03e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.425887] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.493891] env[61867]: DEBUG oslo_concurrency.lockutils [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquiring lock "refresh_cache-f426dd3e-2806-4d70-8784-75c611fd03e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.493891] env[61867]: DEBUG oslo_concurrency.lockutils [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquired lock "refresh_cache-f426dd3e-2806-4d70-8784-75c611fd03e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.498021] env[61867]: DEBUG nova.network.neutron [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 972.674552] env[61867]: DEBUG oslo_vmware.api [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1276986, 'name': PowerOnVM_Task, 'duration_secs': 0.707043} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.675110] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 972.675517] env[61867]: INFO nova.compute.manager [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Took 7.70 seconds to spawn the instance on the hypervisor. [ 972.675885] env[61867]: DEBUG nova.compute.manager [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 972.679369] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e232c1a2-5a1c-4e4b-b0eb-b121f80de4cf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.911570] env[61867]: DEBUG nova.compute.utils [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 972.915190] env[61867]: INFO nova.compute.claims [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 972.918931] env[61867]: DEBUG nova.compute.manager [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 972.920275] env[61867]: DEBUG nova.network.neutron [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 972.960832] env[61867]: DEBUG nova.policy [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '42335b13263a475cbcafd5b041c82cef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2fd4e1da53c34abe8ac2b480f3d7879c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 973.047745] env[61867]: DEBUG nova.network.neutron [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 973.208259] env[61867]: INFO nova.compute.manager [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Took 19.50 seconds to build instance. [ 973.224683] env[61867]: DEBUG nova.network.neutron [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Successfully created port: 213271df-f382-4d6d-b1ec-2077126a4827 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 973.320660] env[61867]: DEBUG nova.compute.manager [req-c18dda6b-f607-4392-b7dc-cb7d8bd4fa7e req-5f71dd9b-5d7a-49db-9d1a-537738f0de3d service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Received event network-vif-plugged-bb767aa8-7bfe-4604-9152-059b0eaa3311 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 973.320660] env[61867]: DEBUG oslo_concurrency.lockutils [req-c18dda6b-f607-4392-b7dc-cb7d8bd4fa7e req-5f71dd9b-5d7a-49db-9d1a-537738f0de3d service nova] Acquiring lock "f426dd3e-2806-4d70-8784-75c611fd03e4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.320803] env[61867]: DEBUG oslo_concurrency.lockutils [req-c18dda6b-f607-4392-b7dc-cb7d8bd4fa7e req-5f71dd9b-5d7a-49db-9d1a-537738f0de3d service nova] Lock "f426dd3e-2806-4d70-8784-75c611fd03e4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.320928] env[61867]: DEBUG oslo_concurrency.lockutils [req-c18dda6b-f607-4392-b7dc-cb7d8bd4fa7e req-5f71dd9b-5d7a-49db-9d1a-537738f0de3d service nova] Lock "f426dd3e-2806-4d70-8784-75c611fd03e4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.321118] env[61867]: DEBUG nova.compute.manager [req-c18dda6b-f607-4392-b7dc-cb7d8bd4fa7e req-5f71dd9b-5d7a-49db-9d1a-537738f0de3d service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] No waiting events found dispatching network-vif-plugged-bb767aa8-7bfe-4604-9152-059b0eaa3311 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 973.321289] env[61867]: WARNING nova.compute.manager [req-c18dda6b-f607-4392-b7dc-cb7d8bd4fa7e req-5f71dd9b-5d7a-49db-9d1a-537738f0de3d service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Received unexpected event network-vif-plugged-bb767aa8-7bfe-4604-9152-059b0eaa3311 for instance with vm_state building and task_state spawning. [ 973.321624] env[61867]: DEBUG nova.compute.manager [req-c18dda6b-f607-4392-b7dc-cb7d8bd4fa7e req-5f71dd9b-5d7a-49db-9d1a-537738f0de3d service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Received event network-changed-bb767aa8-7bfe-4604-9152-059b0eaa3311 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 973.321624] env[61867]: DEBUG nova.compute.manager [req-c18dda6b-f607-4392-b7dc-cb7d8bd4fa7e req-5f71dd9b-5d7a-49db-9d1a-537738f0de3d service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Refreshing instance network info cache due to event network-changed-bb767aa8-7bfe-4604-9152-059b0eaa3311. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 973.321761] env[61867]: DEBUG oslo_concurrency.lockutils [req-c18dda6b-f607-4392-b7dc-cb7d8bd4fa7e req-5f71dd9b-5d7a-49db-9d1a-537738f0de3d service nova] Acquiring lock "refresh_cache-f426dd3e-2806-4d70-8784-75c611fd03e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.419459] env[61867]: DEBUG nova.compute.manager [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 973.424672] env[61867]: INFO nova.compute.resource_tracker [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Updating resource usage from migration 44205de9-d7fc-454a-8217-47bec978b3d8 [ 973.498405] env[61867]: DEBUG nova.network.neutron [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Updating instance_info_cache with network_info: [{"id": "86895cfd-1390-4fcd-921e-e39e1ac94da0", "address": "fa:16:3e:68:e3:8b", "network": {"id": "1c34d845-878d-49cf-aebb-9060726b9ab8", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-352741799", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.171", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "419a5b3f-4c6f-4168-9def-746b4d8c5c24", "external-id": "nsx-vlan-transportzone-656", "segmentation_id": 656, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86895cfd-13", "ovs_interfaceid": "86895cfd-1390-4fcd-921e-e39e1ac94da0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bb767aa8-7bfe-4604-9152-059b0eaa3311", "address": "fa:16:3e:4a:b7:f4", "network": {"id": "51810b4b-272c-45f9-ad6b-efcb0aa08674", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-166211154", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.194", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb767aa8-7b", "ovs_interfaceid": "bb767aa8-7bfe-4604-9152-059b0eaa3311", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.629272] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bebb8c6d-80e8-4a68-a61d-a85576629f46 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.637932] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be576ce-3b6c-4bf9-bfca-007c746aaa2f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.669627] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d6b14f4-5998-4df4-be52-6b50dc1b8129 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.678881] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b721a8-116f-4994-bab7-e4c32ce952ec {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.694781] env[61867]: DEBUG nova.compute.provider_tree [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.710546] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a5aaea7-91f0-44cc-9ee8-4fd70ef899dc tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "0ae83268-ef50-42e6-954d-5e67c4afc20e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.017s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.004880] env[61867]: DEBUG oslo_concurrency.lockutils [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Releasing lock "refresh_cache-f426dd3e-2806-4d70-8784-75c611fd03e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.005028] env[61867]: DEBUG nova.compute.manager [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Instance network_info: |[{"id": "86895cfd-1390-4fcd-921e-e39e1ac94da0", "address": "fa:16:3e:68:e3:8b", "network": {"id": "1c34d845-878d-49cf-aebb-9060726b9ab8", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-352741799", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.171", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "419a5b3f-4c6f-4168-9def-746b4d8c5c24", "external-id": "nsx-vlan-transportzone-656", "segmentation_id": 656, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86895cfd-13", "ovs_interfaceid": "86895cfd-1390-4fcd-921e-e39e1ac94da0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bb767aa8-7bfe-4604-9152-059b0eaa3311", "address": "fa:16:3e:4a:b7:f4", "network": {"id": "51810b4b-272c-45f9-ad6b-efcb0aa08674", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-166211154", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.194", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb767aa8-7b", "ovs_interfaceid": "bb767aa8-7bfe-4604-9152-059b0eaa3311", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 974.005354] env[61867]: DEBUG oslo_concurrency.lockutils [req-c18dda6b-f607-4392-b7dc-cb7d8bd4fa7e req-5f71dd9b-5d7a-49db-9d1a-537738f0de3d service nova] Acquired lock "refresh_cache-f426dd3e-2806-4d70-8784-75c611fd03e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.005542] env[61867]: DEBUG nova.network.neutron [req-c18dda6b-f607-4392-b7dc-cb7d8bd4fa7e req-5f71dd9b-5d7a-49db-9d1a-537738f0de3d service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Refreshing network info cache for port bb767aa8-7bfe-4604-9152-059b0eaa3311 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 974.006742] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:e3:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '419a5b3f-4c6f-4168-9def-746b4d8c5c24', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '86895cfd-1390-4fcd-921e-e39e1ac94da0', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:b7:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a071ecf4-e713-4f97-9271-8c17952f6dee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bb767aa8-7bfe-4604-9152-059b0eaa3311', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 974.015975] env[61867]: DEBUG oslo.service.loopingcall [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 974.016452] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 974.016809] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-58222593-3d17-46d3-8740-58e9704022dc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.044790] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 974.044790] env[61867]: value = "task-1276987" [ 974.044790] env[61867]: _type = "Task" [ 974.044790] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.055586] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276987, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.198309] env[61867]: DEBUG nova.scheduler.client.report [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.432182] env[61867]: DEBUG nova.compute.manager [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 974.455760] env[61867]: DEBUG nova.virt.hardware [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 974.456072] env[61867]: DEBUG nova.virt.hardware [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 974.456243] env[61867]: DEBUG nova.virt.hardware [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 974.456438] env[61867]: DEBUG nova.virt.hardware [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 974.456591] env[61867]: DEBUG nova.virt.hardware [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 974.456742] env[61867]: DEBUG nova.virt.hardware [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 974.456960] env[61867]: DEBUG nova.virt.hardware [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 974.457191] env[61867]: DEBUG nova.virt.hardware [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 974.457360] env[61867]: DEBUG nova.virt.hardware [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 974.457528] env[61867]: DEBUG nova.virt.hardware [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 974.457702] env[61867]: DEBUG nova.virt.hardware [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 974.458600] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2f7b3b3-cbad-4891-b293-319cac9677fa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.467537] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0ac854-35c3-4db4-9e04-46bed1c3bfaa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.558212] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276987, 'name': CreateVM_Task, 'duration_secs': 0.383311} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.558365] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 974.559113] env[61867]: DEBUG oslo_concurrency.lockutils [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.559335] env[61867]: DEBUG oslo_concurrency.lockutils [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.559671] env[61867]: DEBUG oslo_concurrency.lockutils [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 974.559939] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a370a51f-d4e2-4631-a04d-87daccbdbbaa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.565854] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 974.565854] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ab0b97-fa95-5956-fc8e-91298e2c6c3b" [ 974.565854] env[61867]: _type = "Task" [ 974.565854] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.575361] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ab0b97-fa95-5956-fc8e-91298e2c6c3b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.705029] env[61867]: DEBUG oslo_concurrency.lockutils [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.298s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.705029] env[61867]: INFO nova.compute.manager [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Migrating [ 974.712570] env[61867]: DEBUG oslo_concurrency.lockutils [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.020s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.712716] env[61867]: DEBUG nova.objects.instance [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lazy-loading 'resources' on Instance uuid abb41c0c-6d0d-4147-a4af-554ab7d9e921 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.798338] env[61867]: DEBUG nova.network.neutron [req-c18dda6b-f607-4392-b7dc-cb7d8bd4fa7e req-5f71dd9b-5d7a-49db-9d1a-537738f0de3d service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Updated VIF entry in instance network info cache for port bb767aa8-7bfe-4604-9152-059b0eaa3311. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 974.798338] env[61867]: DEBUG nova.network.neutron [req-c18dda6b-f607-4392-b7dc-cb7d8bd4fa7e req-5f71dd9b-5d7a-49db-9d1a-537738f0de3d service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Updating instance_info_cache with network_info: [{"id": "86895cfd-1390-4fcd-921e-e39e1ac94da0", "address": "fa:16:3e:68:e3:8b", "network": {"id": "1c34d845-878d-49cf-aebb-9060726b9ab8", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-352741799", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.171", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "419a5b3f-4c6f-4168-9def-746b4d8c5c24", "external-id": "nsx-vlan-transportzone-656", "segmentation_id": 656, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86895cfd-13", "ovs_interfaceid": "86895cfd-1390-4fcd-921e-e39e1ac94da0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bb767aa8-7bfe-4604-9152-059b0eaa3311", "address": "fa:16:3e:4a:b7:f4", "network": {"id": "51810b4b-272c-45f9-ad6b-efcb0aa08674", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-166211154", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.194", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "60b35760e3e14245aea3600d36c838dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a071ecf4-e713-4f97-9271-8c17952f6dee", "external-id": "nsx-vlan-transportzone-23", "segmentation_id": 23, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb767aa8-7b", "ovs_interfaceid": "bb767aa8-7bfe-4604-9152-059b0eaa3311", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.081074] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ab0b97-fa95-5956-fc8e-91298e2c6c3b, 'name': SearchDatastore_Task, 'duration_secs': 0.0098} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.081074] env[61867]: DEBUG oslo_concurrency.lockutils [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.081074] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 975.081074] env[61867]: DEBUG oslo_concurrency.lockutils [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.081074] env[61867]: DEBUG oslo_concurrency.lockutils [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.081074] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 975.081074] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-90aebef2-456f-4e2b-9311-30dba4b488ea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.096528] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 975.096528] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 975.096947] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eea26d20-61d6-48d1-870d-2d0789706e89 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.105741] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 975.105741] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52dc323f-1eae-f995-6204-abc766109ecb" [ 975.105741] env[61867]: _type = "Task" [ 975.105741] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.118153] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52dc323f-1eae-f995-6204-abc766109ecb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.225578] env[61867]: DEBUG oslo_concurrency.lockutils [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "refresh_cache-6e2136e7-c7e0-4a98-9899-f79c10f0e703" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.225904] env[61867]: DEBUG oslo_concurrency.lockutils [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "refresh_cache-6e2136e7-c7e0-4a98-9899-f79c10f0e703" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.226319] env[61867]: DEBUG nova.network.neutron [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 975.305488] env[61867]: DEBUG oslo_concurrency.lockutils [req-c18dda6b-f607-4392-b7dc-cb7d8bd4fa7e req-5f71dd9b-5d7a-49db-9d1a-537738f0de3d service nova] Releasing lock "refresh_cache-f426dd3e-2806-4d70-8784-75c611fd03e4" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.352591] env[61867]: DEBUG nova.network.neutron [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Successfully updated port: 213271df-f382-4d6d-b1ec-2077126a4827 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 975.371323] env[61867]: DEBUG nova.compute.manager [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Received event network-vif-plugged-213271df-f382-4d6d-b1ec-2077126a4827 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 975.371627] env[61867]: DEBUG oslo_concurrency.lockutils [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] Acquiring lock "d19e7706-e6ac-47dd-8ec3-c30bd315559d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.371861] env[61867]: DEBUG oslo_concurrency.lockutils [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] Lock "d19e7706-e6ac-47dd-8ec3-c30bd315559d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.372033] env[61867]: DEBUG oslo_concurrency.lockutils [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] Lock "d19e7706-e6ac-47dd-8ec3-c30bd315559d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.372228] env[61867]: DEBUG nova.compute.manager [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] No waiting events found dispatching network-vif-plugged-213271df-f382-4d6d-b1ec-2077126a4827 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 975.372413] env[61867]: WARNING nova.compute.manager [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Received unexpected event network-vif-plugged-213271df-f382-4d6d-b1ec-2077126a4827 for instance with vm_state building and task_state spawning. [ 975.372588] env[61867]: DEBUG nova.compute.manager [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Received event network-changed-c05702c4-5138-450d-82b2-790b86d8a59b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 975.372916] env[61867]: DEBUG nova.compute.manager [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Refreshing instance network info cache due to event network-changed-c05702c4-5138-450d-82b2-790b86d8a59b. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 975.372916] env[61867]: DEBUG oslo_concurrency.lockutils [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] Acquiring lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.373079] env[61867]: DEBUG oslo_concurrency.lockutils [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] Acquired lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.373233] env[61867]: DEBUG nova.network.neutron [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Refreshing network info cache for port c05702c4-5138-450d-82b2-790b86d8a59b {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 975.480496] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da552094-dc21-4f03-9cbe-bc6266f7f682 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.488981] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af1f3199-1406-4ef7-acae-921210d70490 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.521126] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d8e797-6fe9-4d4a-bd81-9cc2f5fd4c40 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.529132] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d086604-3f07-471c-bb3c-b8385271a528 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.542817] env[61867]: DEBUG nova.compute.provider_tree [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.618602] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52dc323f-1eae-f995-6204-abc766109ecb, 'name': SearchDatastore_Task, 'duration_secs': 0.009496} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.619423] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d310e2b-b1da-40e0-a035-1e597abdabe5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.625134] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 975.625134] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522cbbdb-84d3-8be5-5e22-4447cc2ef038" [ 975.625134] env[61867]: _type = "Task" [ 975.625134] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.633801] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522cbbdb-84d3-8be5-5e22-4447cc2ef038, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.854259] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Acquiring lock "refresh_cache-d19e7706-e6ac-47dd-8ec3-c30bd315559d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.854259] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Acquired lock "refresh_cache-d19e7706-e6ac-47dd-8ec3-c30bd315559d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.854465] env[61867]: DEBUG nova.network.neutron [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 976.045386] env[61867]: DEBUG nova.scheduler.client.report [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.138862] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522cbbdb-84d3-8be5-5e22-4447cc2ef038, 'name': SearchDatastore_Task, 'duration_secs': 0.010598} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.138862] env[61867]: DEBUG oslo_concurrency.lockutils [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.138862] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] f426dd3e-2806-4d70-8784-75c611fd03e4/f426dd3e-2806-4d70-8784-75c611fd03e4.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 976.138862] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a21ec5d1-894f-4ef0-9e42-f255ff50b355 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.145802] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 976.145802] env[61867]: value = "task-1276988" [ 976.145802] env[61867]: _type = "Task" [ 976.145802] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.154058] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276988, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.166187] env[61867]: DEBUG nova.network.neutron [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Updating instance_info_cache with network_info: [{"id": "e070a089-e0e7-4b11-9688-26d528b37e6a", "address": "fa:16:3e:7b:ac:a0", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape070a089-e0", "ovs_interfaceid": "e070a089-e0e7-4b11-9688-26d528b37e6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.175108] env[61867]: DEBUG nova.network.neutron [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Updated VIF entry in instance network info cache for port c05702c4-5138-450d-82b2-790b86d8a59b. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 976.175506] env[61867]: DEBUG nova.network.neutron [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Updating instance_info_cache with network_info: [{"id": "c05702c4-5138-450d-82b2-790b86d8a59b", "address": "fa:16:3e:cf:d4:2a", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc05702c4-51", "ovs_interfaceid": "c05702c4-5138-450d-82b2-790b86d8a59b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.392897] env[61867]: DEBUG nova.network.neutron [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 976.542326] env[61867]: DEBUG nova.network.neutron [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Updating instance_info_cache with network_info: [{"id": "213271df-f382-4d6d-b1ec-2077126a4827", "address": "fa:16:3e:e5:e1:e3", "network": {"id": "88b7d521-ebfc-4432-89a3-ea72653b40fd", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1053617647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fd4e1da53c34abe8ac2b480f3d7879c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap213271df-f3", "ovs_interfaceid": "213271df-f382-4d6d-b1ec-2077126a4827", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.554302] env[61867]: DEBUG oslo_concurrency.lockutils [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.842s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.558319] env[61867]: DEBUG oslo_concurrency.lockutils [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.631s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.560125] env[61867]: INFO nova.compute.claims [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 976.585689] env[61867]: INFO nova.scheduler.client.report [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Deleted allocations for instance abb41c0c-6d0d-4147-a4af-554ab7d9e921 [ 976.657051] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276988, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.672123] env[61867]: DEBUG oslo_concurrency.lockutils [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "refresh_cache-6e2136e7-c7e0-4a98-9899-f79c10f0e703" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.679211] env[61867]: DEBUG oslo_concurrency.lockutils [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] Releasing lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.679530] env[61867]: DEBUG nova.compute.manager [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Received event network-changed-a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.679750] env[61867]: DEBUG nova.compute.manager [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Refreshing instance network info cache due to event network-changed-a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 976.680014] env[61867]: DEBUG oslo_concurrency.lockutils [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] Acquiring lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.680230] env[61867]: DEBUG oslo_concurrency.lockutils [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] Acquired lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.680537] env[61867]: DEBUG nova.network.neutron [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Refreshing network info cache for port a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 977.046055] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Releasing lock "refresh_cache-d19e7706-e6ac-47dd-8ec3-c30bd315559d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.046214] env[61867]: DEBUG nova.compute.manager [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Instance network_info: |[{"id": "213271df-f382-4d6d-b1ec-2077126a4827", "address": "fa:16:3e:e5:e1:e3", "network": {"id": "88b7d521-ebfc-4432-89a3-ea72653b40fd", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1053617647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fd4e1da53c34abe8ac2b480f3d7879c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap213271df-f3", "ovs_interfaceid": "213271df-f382-4d6d-b1ec-2077126a4827", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 977.046659] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e5:e1:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4223acd2-30f7-440e-b975-60b30d931694', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '213271df-f382-4d6d-b1ec-2077126a4827', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 977.054339] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Creating folder: Project (2fd4e1da53c34abe8ac2b480f3d7879c). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 977.054665] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-81d93291-c6a0-48c1-80ef-925f2b388847 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.070965] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Created folder: Project (2fd4e1da53c34abe8ac2b480f3d7879c) in parent group-v274258. [ 977.071323] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Creating folder: Instances. Parent ref: group-v274407. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 977.071831] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-466b6d3b-d0bd-443b-b512-fcd59a76e7c8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.084269] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Created folder: Instances in parent group-v274407. [ 977.084560] env[61867]: DEBUG oslo.service.loopingcall [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 977.084789] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 977.085060] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fadb29f8-6251-4b92-ae9a-d218302d5784 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.104446] env[61867]: DEBUG oslo_concurrency.lockutils [None req-961b6ae6-2837-4a62-a4de-de0e6d416140 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "abb41c0c-6d0d-4147-a4af-554ab7d9e921" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.808s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.112219] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 977.112219] env[61867]: value = "task-1276991" [ 977.112219] env[61867]: _type = "Task" [ 977.112219] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.120649] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276991, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.159842] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276988, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516337} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.159842] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] f426dd3e-2806-4d70-8784-75c611fd03e4/f426dd3e-2806-4d70-8784-75c611fd03e4.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 977.159842] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 977.160388] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f937ff0-2d71-48fc-bf94-b197075aede2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.169022] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 977.169022] env[61867]: value = "task-1276992" [ 977.169022] env[61867]: _type = "Task" [ 977.169022] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.180654] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276992, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.389565] env[61867]: DEBUG nova.network.neutron [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Updated VIF entry in instance network info cache for port a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 977.389824] env[61867]: DEBUG nova.network.neutron [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Updating instance_info_cache with network_info: [{"id": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "address": "fa:16:3e:a0:23:06", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3537cac-b5", "ovs_interfaceid": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.440158] env[61867]: DEBUG nova.compute.manager [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Received event network-changed-213271df-f382-4d6d-b1ec-2077126a4827 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 977.440278] env[61867]: DEBUG nova.compute.manager [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Refreshing instance network info cache due to event network-changed-213271df-f382-4d6d-b1ec-2077126a4827. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 977.440541] env[61867]: DEBUG oslo_concurrency.lockutils [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] Acquiring lock "refresh_cache-d19e7706-e6ac-47dd-8ec3-c30bd315559d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.440680] env[61867]: DEBUG oslo_concurrency.lockutils [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] Acquired lock "refresh_cache-d19e7706-e6ac-47dd-8ec3-c30bd315559d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.440868] env[61867]: DEBUG nova.network.neutron [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Refreshing network info cache for port 213271df-f382-4d6d-b1ec-2077126a4827 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 977.621625] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276991, 'name': CreateVM_Task, 'duration_secs': 0.388287} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.621928] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 977.622817] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.623058] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.623532] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 977.623832] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3099d08c-eff5-469c-b843-d8e8603c42e6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.631818] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Waiting for the task: (returnval){ [ 977.631818] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52180619-94fb-3761-696d-7a778b9b8ed9" [ 977.631818] env[61867]: _type = "Task" [ 977.631818] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.644315] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52180619-94fb-3761-696d-7a778b9b8ed9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.681218] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276992, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.11036} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.683834] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 977.687718] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b5f6f9-10c3-4c74-b595-e96c9409aa25 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.715624] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] f426dd3e-2806-4d70-8784-75c611fd03e4/f426dd3e-2806-4d70-8784-75c611fd03e4.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 977.719193] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e46841af-0f9a-4e26-82a2-465d40e043d1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.742353] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 977.742353] env[61867]: value = "task-1276993" [ 977.742353] env[61867]: _type = "Task" [ 977.742353] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.748290] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "7479bf91-5aef-4e75-a127-7e82ae15a003" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.748575] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "7479bf91-5aef-4e75-a127-7e82ae15a003" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.756663] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276993, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.892056] env[61867]: DEBUG oslo_concurrency.lockutils [req-4768985e-ca0e-41f8-bd59-2561bc2c6de4 req-9dab0620-691e-45a0-82f0-761724ab1edf service nova] Releasing lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.893190] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c317616-4ef4-494a-87bd-b3e5283206c8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.901708] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2ef6a8-5ee7-4b99-a9c8-782ce3426d37 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.932844] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1afa1ad-3710-4be7-983c-9a51a053b58f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.941547] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc30c79f-1d1d-446a-84ac-47791d8e0ed3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.957839] env[61867]: DEBUG nova.compute.provider_tree [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.001083] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.001341] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.144074] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52180619-94fb-3761-696d-7a778b9b8ed9, 'name': SearchDatastore_Task, 'duration_secs': 0.010576} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.144415] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.144769] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 978.144951] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.145035] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.145221] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 978.145477] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c288b10-bbc0-4bcc-a152-a3dbd4d6d95a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.153747] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 978.153934] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 978.154640] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4dbb9d4-92aa-4fb0-9bc8-a7a26a7d5a11 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.159840] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Waiting for the task: (returnval){ [ 978.159840] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5220569e-e827-769f-2cdd-6c8a3a2e43bc" [ 978.159840] env[61867]: _type = "Task" [ 978.159840] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.167373] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5220569e-e827-769f-2cdd-6c8a3a2e43bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.184414] env[61867]: DEBUG nova.network.neutron [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Updated VIF entry in instance network info cache for port 213271df-f382-4d6d-b1ec-2077126a4827. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 978.184793] env[61867]: DEBUG nova.network.neutron [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Updating instance_info_cache with network_info: [{"id": "213271df-f382-4d6d-b1ec-2077126a4827", "address": "fa:16:3e:e5:e1:e3", "network": {"id": "88b7d521-ebfc-4432-89a3-ea72653b40fd", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1053617647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fd4e1da53c34abe8ac2b480f3d7879c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap213271df-f3", "ovs_interfaceid": "213271df-f382-4d6d-b1ec-2077126a4827", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.194680] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1a164e-7f95-4bf2-b311-41863aed7f13 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.213283] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Updating instance '6e2136e7-c7e0-4a98-9899-f79c10f0e703' progress to 0 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 978.252773] env[61867]: INFO nova.compute.manager [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Detaching volume 9bcdabad-faa3-4b43-ac62-c59cb70802e4 [ 978.254401] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276993, 'name': ReconfigVM_Task, 'duration_secs': 0.296166} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.255473] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Reconfigured VM instance instance-00000060 to attach disk [datastore2] f426dd3e-2806-4d70-8784-75c611fd03e4/f426dd3e-2806-4d70-8784-75c611fd03e4.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.256248] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f288f59d-b2f4-4cf1-ac0a-3a87ecde82c3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.264029] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 978.264029] env[61867]: value = "task-1276994" [ 978.264029] env[61867]: _type = "Task" [ 978.264029] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.272824] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276994, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.289530] env[61867]: INFO nova.virt.block_device [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Attempting to driver detach volume 9bcdabad-faa3-4b43-ac62-c59cb70802e4 from mountpoint /dev/sdb [ 978.289775] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Volume detach. Driver type: vmdk {{(pid=61867) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 978.289967] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274394', 'volume_id': '9bcdabad-faa3-4b43-ac62-c59cb70802e4', 'name': 'volume-9bcdabad-faa3-4b43-ac62-c59cb70802e4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7479bf91-5aef-4e75-a127-7e82ae15a003', 'attached_at': '', 'detached_at': '', 'volume_id': '9bcdabad-faa3-4b43-ac62-c59cb70802e4', 'serial': '9bcdabad-faa3-4b43-ac62-c59cb70802e4'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 978.291011] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef732b05-0c9e-4d16-b0c0-82c57218f2df {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.313056] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a317838-06e8-402a-a6bd-f32e0cc46d13 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.320147] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-744bda8a-bd4b-412c-a32c-97c6cf2dc92a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.341044] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a480c3-1da3-4bc3-8b1c-4cf4bda9162f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.357546] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] The volume has not been displaced from its original location: [datastore2] volume-9bcdabad-faa3-4b43-ac62-c59cb70802e4/volume-9bcdabad-faa3-4b43-ac62-c59cb70802e4.vmdk. No consolidation needed. {{(pid=61867) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 978.362751] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Reconfiguring VM instance instance-00000050 to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 978.363069] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a23025aa-325b-4a1f-8919-c047d5ff4de4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.382804] env[61867]: DEBUG oslo_vmware.api [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 978.382804] env[61867]: value = "task-1276995" [ 978.382804] env[61867]: _type = "Task" [ 978.382804] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.391058] env[61867]: DEBUG oslo_vmware.api [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276995, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.461912] env[61867]: DEBUG nova.scheduler.client.report [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 978.504082] env[61867]: DEBUG nova.compute.manager [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 978.670362] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5220569e-e827-769f-2cdd-6c8a3a2e43bc, 'name': SearchDatastore_Task, 'duration_secs': 0.008879} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.671513] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2aa9eba-7a90-48b0-b7b1-efd9ea643c13 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.677334] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Waiting for the task: (returnval){ [ 978.677334] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5231e708-102c-7d05-ef9c-82f8fea0a0d7" [ 978.677334] env[61867]: _type = "Task" [ 978.677334] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.685134] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5231e708-102c-7d05-ef9c-82f8fea0a0d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.687691] env[61867]: DEBUG oslo_concurrency.lockutils [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] Releasing lock "refresh_cache-d19e7706-e6ac-47dd-8ec3-c30bd315559d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.687942] env[61867]: DEBUG nova.compute.manager [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Received event network-changed-a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.688136] env[61867]: DEBUG nova.compute.manager [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Refreshing instance network info cache due to event network-changed-a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 978.688348] env[61867]: DEBUG oslo_concurrency.lockutils [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] Acquiring lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.688498] env[61867]: DEBUG oslo_concurrency.lockutils [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] Acquired lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.688660] env[61867]: DEBUG nova.network.neutron [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Refreshing network info cache for port a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 978.719014] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 978.719308] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e8194b8-1dde-4710-905e-5248c629a397 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.726930] env[61867]: DEBUG oslo_vmware.api [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 978.726930] env[61867]: value = "task-1276996" [ 978.726930] env[61867]: _type = "Task" [ 978.726930] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.735518] env[61867]: DEBUG oslo_vmware.api [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276996, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.774543] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276994, 'name': Rename_Task, 'duration_secs': 0.151028} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.774837] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 978.775104] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f95c706-39fa-4e57-84ca-839e14515dbe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.782682] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 978.782682] env[61867]: value = "task-1276997" [ 978.782682] env[61867]: _type = "Task" [ 978.782682] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.790750] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276997, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.893154] env[61867]: DEBUG oslo_vmware.api [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276995, 'name': ReconfigVM_Task, 'duration_secs': 0.235804} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.893497] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Reconfigured VM instance instance-00000050 to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 978.898301] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b8079fe-6f7a-44b6-a0cf-eae7891147d4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.913662] env[61867]: DEBUG oslo_vmware.api [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 978.913662] env[61867]: value = "task-1276998" [ 978.913662] env[61867]: _type = "Task" [ 978.913662] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.922185] env[61867]: DEBUG oslo_vmware.api [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276998, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.966874] env[61867]: DEBUG oslo_concurrency.lockutils [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.967419] env[61867]: DEBUG nova.compute.manager [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 978.970490] env[61867]: DEBUG oslo_concurrency.lockutils [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.953s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.971634] env[61867]: DEBUG nova.objects.instance [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lazy-loading 'resources' on Instance uuid 8a7f4314-0fd4-49f9-8eb6-12baa0977a53 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.028097] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.193399] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5231e708-102c-7d05-ef9c-82f8fea0a0d7, 'name': SearchDatastore_Task, 'duration_secs': 0.011089} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.193734] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.193955] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] d19e7706-e6ac-47dd-8ec3-c30bd315559d/d19e7706-e6ac-47dd-8ec3-c30bd315559d.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 979.194591] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8c612348-4b7b-407a-b8e9-44e9af357447 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.201959] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Waiting for the task: (returnval){ [ 979.201959] env[61867]: value = "task-1276999" [ 979.201959] env[61867]: _type = "Task" [ 979.201959] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.209930] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1276999, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.237747] env[61867]: DEBUG oslo_vmware.api [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1276996, 'name': PowerOffVM_Task, 'duration_secs': 0.19029} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.238046] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 979.238243] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Updating instance '6e2136e7-c7e0-4a98-9899-f79c10f0e703' progress to 17 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 979.292940] env[61867]: DEBUG oslo_vmware.api [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1276997, 'name': PowerOnVM_Task, 'duration_secs': 0.457315} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.293487] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 979.293731] env[61867]: INFO nova.compute.manager [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Took 10.99 seconds to spawn the instance on the hypervisor. [ 979.293925] env[61867]: DEBUG nova.compute.manager [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 979.295324] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32016917-7193-4430-b452-29daffa915cb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.402770] env[61867]: DEBUG nova.network.neutron [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Updated VIF entry in instance network info cache for port a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 979.403187] env[61867]: DEBUG nova.network.neutron [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Updating instance_info_cache with network_info: [{"id": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "address": "fa:16:3e:a0:23:06", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3537cac-b5", "ovs_interfaceid": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.426245] env[61867]: DEBUG oslo_vmware.api [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1276998, 'name': ReconfigVM_Task, 'duration_secs': 0.141512} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.426245] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274394', 'volume_id': '9bcdabad-faa3-4b43-ac62-c59cb70802e4', 'name': 'volume-9bcdabad-faa3-4b43-ac62-c59cb70802e4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7479bf91-5aef-4e75-a127-7e82ae15a003', 'attached_at': '', 'detached_at': '', 'volume_id': '9bcdabad-faa3-4b43-ac62-c59cb70802e4', 'serial': '9bcdabad-faa3-4b43-ac62-c59cb70802e4'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 979.474088] env[61867]: DEBUG nova.compute.utils [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 979.477875] env[61867]: DEBUG nova.compute.manager [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 979.478077] env[61867]: DEBUG nova.network.neutron [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 979.525425] env[61867]: DEBUG nova.policy [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'adcd9eb75ecc4eccb335ebfec207a900', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7316e4f263a9432ab2f9f91484d62f58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 979.683494] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e91d25cb-c945-4589-a9bc-a3485397a369 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.694938] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7beb92a-9550-4c95-b7f1-a1931577fc14 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.730722] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e36e39-4c96-4ed7-9e80-6419c7d4360e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.736652] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1276999, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.741761] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e98860-ef6a-480c-ae5f-98488ec2a250 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.749520] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 979.749771] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 979.749914] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.750113] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 979.750309] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.750489] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 979.750714] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 979.750924] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 979.751172] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 979.751364] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 979.751543] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 979.756880] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd110b1a-1ced-4865-ab6e-d096088d3111 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.777963] env[61867]: DEBUG nova.compute.provider_tree [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 979.784555] env[61867]: DEBUG oslo_vmware.api [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 979.784555] env[61867]: value = "task-1277000" [ 979.784555] env[61867]: _type = "Task" [ 979.784555] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.795489] env[61867]: DEBUG oslo_vmware.api [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277000, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.814229] env[61867]: DEBUG nova.network.neutron [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Successfully created port: 92925b9b-00ea-4282-9781-7e60212ae093 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 979.818519] env[61867]: INFO nova.compute.manager [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Took 24.59 seconds to build instance. [ 979.906121] env[61867]: DEBUG oslo_concurrency.lockutils [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] Releasing lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.906319] env[61867]: DEBUG nova.compute.manager [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Received event network-changed-c05702c4-5138-450d-82b2-790b86d8a59b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 979.906319] env[61867]: DEBUG nova.compute.manager [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Refreshing instance network info cache due to event network-changed-c05702c4-5138-450d-82b2-790b86d8a59b. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 979.906567] env[61867]: DEBUG oslo_concurrency.lockutils [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] Acquiring lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.906720] env[61867]: DEBUG oslo_concurrency.lockutils [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] Acquired lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.906885] env[61867]: DEBUG nova.network.neutron [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Refreshing network info cache for port c05702c4-5138-450d-82b2-790b86d8a59b {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 979.974644] env[61867]: DEBUG nova.objects.instance [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lazy-loading 'flavor' on Instance uuid 7479bf91-5aef-4e75-a127-7e82ae15a003 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.982017] env[61867]: DEBUG nova.compute.manager [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 980.213991] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1276999, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.536161} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.214285] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] d19e7706-e6ac-47dd-8ec3-c30bd315559d/d19e7706-e6ac-47dd-8ec3-c30bd315559d.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 980.214515] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 980.214768] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd5f4328-49ed-4ad3-aa5a-a438f4a09f70 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.221795] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Waiting for the task: (returnval){ [ 980.221795] env[61867]: value = "task-1277001" [ 980.221795] env[61867]: _type = "Task" [ 980.221795] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.229448] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1277001, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.286887] env[61867]: DEBUG nova.scheduler.client.report [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.299595] env[61867]: DEBUG oslo_vmware.api [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277000, 'name': ReconfigVM_Task, 'duration_secs': 0.277824} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.299903] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Updating instance '6e2136e7-c7e0-4a98-9899-f79c10f0e703' progress to 33 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 980.320341] env[61867]: DEBUG oslo_concurrency.lockutils [None req-98375855-9edd-4f0f-b632-283ffa928bf9 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "f426dd3e-2806-4d70-8784-75c611fd03e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.105s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.639982] env[61867]: DEBUG nova.network.neutron [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Updated VIF entry in instance network info cache for port c05702c4-5138-450d-82b2-790b86d8a59b. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 980.640518] env[61867]: DEBUG nova.network.neutron [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Updating instance_info_cache with network_info: [{"id": "c05702c4-5138-450d-82b2-790b86d8a59b", "address": "fa:16:3e:cf:d4:2a", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc05702c4-51", "ovs_interfaceid": "c05702c4-5138-450d-82b2-790b86d8a59b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.646702] env[61867]: DEBUG oslo_concurrency.lockutils [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquiring lock "f426dd3e-2806-4d70-8784-75c611fd03e4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.646947] env[61867]: DEBUG oslo_concurrency.lockutils [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "f426dd3e-2806-4d70-8784-75c611fd03e4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.647173] env[61867]: DEBUG oslo_concurrency.lockutils [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquiring lock "f426dd3e-2806-4d70-8784-75c611fd03e4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.647365] env[61867]: DEBUG oslo_concurrency.lockutils [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "f426dd3e-2806-4d70-8784-75c611fd03e4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.647542] env[61867]: DEBUG oslo_concurrency.lockutils [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "f426dd3e-2806-4d70-8784-75c611fd03e4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.649551] env[61867]: INFO nova.compute.manager [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Terminating instance [ 980.651479] env[61867]: DEBUG nova.compute.manager [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 980.651684] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 980.653276] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b98778c8-e120-4949-b267-7c68512fbe2a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.661767] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 980.662026] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b234e041-7daa-4619-b9de-f6278a46c80f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.673800] env[61867]: DEBUG oslo_vmware.api [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 980.673800] env[61867]: value = "task-1277002" [ 980.673800] env[61867]: _type = "Task" [ 980.673800] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.683143] env[61867]: DEBUG oslo_vmware.api [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1277002, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.732229] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1277001, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072332} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.732569] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 980.733347] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ae4727-daa0-429f-8720-1a6a427657df {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.756214] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] d19e7706-e6ac-47dd-8ec3-c30bd315559d/d19e7706-e6ac-47dd-8ec3-c30bd315559d.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 980.756843] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4fcbc17b-81af-459f-8bc2-2a13d5d2fe59 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.777939] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Waiting for the task: (returnval){ [ 980.777939] env[61867]: value = "task-1277003" [ 980.777939] env[61867]: _type = "Task" [ 980.777939] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.787885] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1277003, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.794854] env[61867]: DEBUG oslo_concurrency.lockutils [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.824s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.797982] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.371s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.801678] env[61867]: INFO nova.compute.claims [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 980.806690] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 980.806958] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 980.807121] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 980.807323] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 980.807472] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 980.808215] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 980.808215] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 980.808448] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 980.808716] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 980.809295] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 980.809295] env[61867]: DEBUG nova.virt.hardware [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 980.815253] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Reconfiguring VM instance instance-0000005e to detach disk 2000 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 980.815926] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-692ad769-c88b-4a8e-8b55-914b92dc150d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.831583] env[61867]: INFO nova.scheduler.client.report [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Deleted allocations for instance 8a7f4314-0fd4-49f9-8eb6-12baa0977a53 [ 980.842365] env[61867]: DEBUG oslo_vmware.api [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 980.842365] env[61867]: value = "task-1277004" [ 980.842365] env[61867]: _type = "Task" [ 980.842365] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.851866] env[61867]: DEBUG oslo_vmware.api [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277004, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.982636] env[61867]: DEBUG oslo_concurrency.lockutils [None req-54a24054-f470-4b87-b34b-2ec74e9f34b8 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "7479bf91-5aef-4e75-a127-7e82ae15a003" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.234s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.990464] env[61867]: DEBUG nova.compute.manager [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 981.016955] env[61867]: DEBUG nova.virt.hardware [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 981.017262] env[61867]: DEBUG nova.virt.hardware [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 981.017460] env[61867]: DEBUG nova.virt.hardware [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 981.017689] env[61867]: DEBUG nova.virt.hardware [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 981.017860] env[61867]: DEBUG nova.virt.hardware [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 981.018036] env[61867]: DEBUG nova.virt.hardware [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 981.018263] env[61867]: DEBUG nova.virt.hardware [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 981.018456] env[61867]: DEBUG nova.virt.hardware [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 981.018710] env[61867]: DEBUG nova.virt.hardware [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 981.018905] env[61867]: DEBUG nova.virt.hardware [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 981.019125] env[61867]: DEBUG nova.virt.hardware [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 981.020035] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfed7246-42b0-48c9-bf9b-28f8121f3f4c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.028210] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc91565-7448-4687-8122-43d5a64045a2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.143600] env[61867]: DEBUG oslo_concurrency.lockutils [req-a9ca2a05-e88c-4906-9ae7-b8a29f03ae7c req-b927ad6c-54ce-4ba4-9b42-8bfd339d21e2 service nova] Releasing lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.172981] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "7479bf91-5aef-4e75-a127-7e82ae15a003" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.173235] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "7479bf91-5aef-4e75-a127-7e82ae15a003" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.173500] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "7479bf91-5aef-4e75-a127-7e82ae15a003-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.173700] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "7479bf91-5aef-4e75-a127-7e82ae15a003-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.173878] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "7479bf91-5aef-4e75-a127-7e82ae15a003-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.178959] env[61867]: INFO nova.compute.manager [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Terminating instance [ 981.181714] env[61867]: DEBUG nova.compute.manager [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 981.181714] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 981.182259] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82d250b-d649-4b0f-8997-77c376c62680 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.188163] env[61867]: DEBUG oslo_vmware.api [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1277002, 'name': PowerOffVM_Task, 'duration_secs': 0.224362} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.188755] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 981.188925] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 981.189196] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5acec2a6-562f-46bf-97da-5cb555655e88 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.192815] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 981.193049] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5fcddd1b-1c9a-4626-9153-e6057acc75a0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.200506] env[61867]: DEBUG oslo_vmware.api [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 981.200506] env[61867]: value = "task-1277006" [ 981.200506] env[61867]: _type = "Task" [ 981.200506] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.209203] env[61867]: DEBUG oslo_vmware.api [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1277006, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.271132] env[61867]: DEBUG nova.compute.manager [req-1012879f-78b4-45b5-ab5e-fb33bd853d50 req-23bfa850-77d2-4955-a16e-b04483306a30 service nova] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Received event network-vif-plugged-92925b9b-00ea-4282-9781-7e60212ae093 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.271383] env[61867]: DEBUG oslo_concurrency.lockutils [req-1012879f-78b4-45b5-ab5e-fb33bd853d50 req-23bfa850-77d2-4955-a16e-b04483306a30 service nova] Acquiring lock "976299f8-c6cb-4106-90f2-0a4c85625d8c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.271653] env[61867]: DEBUG oslo_concurrency.lockutils [req-1012879f-78b4-45b5-ab5e-fb33bd853d50 req-23bfa850-77d2-4955-a16e-b04483306a30 service nova] Lock "976299f8-c6cb-4106-90f2-0a4c85625d8c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.271957] env[61867]: DEBUG oslo_concurrency.lockutils [req-1012879f-78b4-45b5-ab5e-fb33bd853d50 req-23bfa850-77d2-4955-a16e-b04483306a30 service nova] Lock "976299f8-c6cb-4106-90f2-0a4c85625d8c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.272093] env[61867]: DEBUG nova.compute.manager [req-1012879f-78b4-45b5-ab5e-fb33bd853d50 req-23bfa850-77d2-4955-a16e-b04483306a30 service nova] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] No waiting events found dispatching network-vif-plugged-92925b9b-00ea-4282-9781-7e60212ae093 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 981.272273] env[61867]: WARNING nova.compute.manager [req-1012879f-78b4-45b5-ab5e-fb33bd853d50 req-23bfa850-77d2-4955-a16e-b04483306a30 service nova] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Received unexpected event network-vif-plugged-92925b9b-00ea-4282-9781-7e60212ae093 for instance with vm_state building and task_state spawning. [ 981.287932] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1277003, 'name': ReconfigVM_Task, 'duration_secs': 0.311326} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.288223] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Reconfigured VM instance instance-00000061 to attach disk [datastore1] d19e7706-e6ac-47dd-8ec3-c30bd315559d/d19e7706-e6ac-47dd-8ec3-c30bd315559d.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 981.288857] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bb0a9d7c-4edd-4ea1-9719-bfaf59a77280 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.297512] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Waiting for the task: (returnval){ [ 981.297512] env[61867]: value = "task-1277007" [ 981.297512] env[61867]: _type = "Task" [ 981.297512] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.309216] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1277007, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.314557] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 981.314890] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 981.315036] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Deleting the datastore file [datastore2] f426dd3e-2806-4d70-8784-75c611fd03e4 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.316148] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-03ec8c45-b2c3-470f-85ea-239410cef421 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.323019] env[61867]: DEBUG oslo_vmware.api [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for the task: (returnval){ [ 981.323019] env[61867]: value = "task-1277008" [ 981.323019] env[61867]: _type = "Task" [ 981.323019] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.331915] env[61867]: DEBUG oslo_vmware.api [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1277008, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.344168] env[61867]: DEBUG oslo_concurrency.lockutils [None req-737ebba5-46b0-4972-a631-8e9dc3d60b8e tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "8a7f4314-0fd4-49f9-8eb6-12baa0977a53" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.575s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.357529] env[61867]: DEBUG oslo_vmware.api [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277004, 'name': ReconfigVM_Task, 'duration_secs': 0.181375} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.357873] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Reconfigured VM instance instance-0000005e to detach disk 2000 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 981.359318] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41cc113-1e68-45c9-91a3-96ecd56ced65 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.389181] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 6e2136e7-c7e0-4a98-9899-f79c10f0e703/6e2136e7-c7e0-4a98-9899-f79c10f0e703.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 981.389915] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5753400d-01aa-4b80-977e-090415e2bd76 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.405590] env[61867]: DEBUG nova.network.neutron [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Successfully updated port: 92925b9b-00ea-4282-9781-7e60212ae093 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 981.414395] env[61867]: DEBUG oslo_vmware.api [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 981.414395] env[61867]: value = "task-1277009" [ 981.414395] env[61867]: _type = "Task" [ 981.414395] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.426646] env[61867]: DEBUG oslo_vmware.api [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277009, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.710778] env[61867]: DEBUG oslo_vmware.api [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1277006, 'name': PowerOffVM_Task, 'duration_secs': 0.24964} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.711074] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 981.711271] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 981.711556] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a22f0b07-c5b0-4462-a39b-a1eadeab9ffa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.778423] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 981.778624] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 981.778751] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Deleting the datastore file [datastore1] 7479bf91-5aef-4e75-a127-7e82ae15a003 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.779055] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94a8b1e3-8f58-4ea0-ba54-942dc1fa5623 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.788378] env[61867]: DEBUG oslo_vmware.api [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 981.788378] env[61867]: value = "task-1277011" [ 981.788378] env[61867]: _type = "Task" [ 981.788378] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.797907] env[61867]: DEBUG oslo_vmware.api [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1277011, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.806939] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1277007, 'name': Rename_Task, 'duration_secs': 0.166595} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.807388] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 981.807698] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bc6ade25-a27a-49ce-aad3-533152b44760 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.814774] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Waiting for the task: (returnval){ [ 981.814774] env[61867]: value = "task-1277012" [ 981.814774] env[61867]: _type = "Task" [ 981.814774] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.822933] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1277012, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.834688] env[61867]: DEBUG oslo_vmware.api [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Task: {'id': task-1277008, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193268} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.834956] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 981.835219] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 981.835354] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 981.835532] env[61867]: INFO nova.compute.manager [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Took 1.18 seconds to destroy the instance on the hypervisor. [ 981.835774] env[61867]: DEBUG oslo.service.loopingcall [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.835966] env[61867]: DEBUG nova.compute.manager [-] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 981.836070] env[61867]: DEBUG nova.network.neutron [-] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 981.909327] env[61867]: DEBUG oslo_concurrency.lockutils [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "refresh_cache-976299f8-c6cb-4106-90f2-0a4c85625d8c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.909327] env[61867]: DEBUG oslo_concurrency.lockutils [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "refresh_cache-976299f8-c6cb-4106-90f2-0a4c85625d8c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.909481] env[61867]: DEBUG nova.network.neutron [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 981.928090] env[61867]: DEBUG oslo_vmware.api [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277009, 'name': ReconfigVM_Task, 'duration_secs': 0.292187} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.930792] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 6e2136e7-c7e0-4a98-9899-f79c10f0e703/6e2136e7-c7e0-4a98-9899-f79c10f0e703.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 981.931189] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Updating instance '6e2136e7-c7e0-4a98-9899-f79c10f0e703' progress to 50 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 982.014191] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a6d7bc-6127-4065-beef-e9f7cf73c0a3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.021953] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e341f2-8d01-44d8-a599-4a85548236fd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.051928] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ecdacf9-5d1f-4d8f-a13e-5dca21b1177b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.059474] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b39e8a-dc0e-44e8-b8f6-d74d0d6d72cf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.074690] env[61867]: DEBUG nova.compute.provider_tree [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.299101] env[61867]: DEBUG oslo_vmware.api [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1277011, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150934} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.299382] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 982.299580] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 982.299765] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 982.299946] env[61867]: INFO nova.compute.manager [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Took 1.12 seconds to destroy the instance on the hypervisor. [ 982.300224] env[61867]: DEBUG oslo.service.loopingcall [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.300471] env[61867]: DEBUG nova.compute.manager [-] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 982.300578] env[61867]: DEBUG nova.network.neutron [-] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 982.325240] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1277012, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.441147] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d74413-70f3-4aa2-a0fd-b044263f98d6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.446051] env[61867]: DEBUG nova.network.neutron [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 982.448559] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.448702] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.471276] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d598258-4899-4750-bf61-fc9873b53e83 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.491914] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Updating instance '6e2136e7-c7e0-4a98-9899-f79c10f0e703' progress to 67 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 982.578101] env[61867]: DEBUG nova.scheduler.client.report [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.600554] env[61867]: DEBUG nova.network.neutron [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Updating instance_info_cache with network_info: [{"id": "92925b9b-00ea-4282-9781-7e60212ae093", "address": "fa:16:3e:dd:2b:fa", "network": {"id": "14cee476-296a-4093-99ba-c2a5ef63f7cf", "bridge": "br-int", "label": "tempest-ServersTestJSON-243183479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7316e4f263a9432ab2f9f91484d62f58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92925b9b-00", "ovs_interfaceid": "92925b9b-00ea-4282-9781-7e60212ae093", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.828976] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1277012, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.839522] env[61867]: DEBUG nova.network.neutron [-] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.954160] env[61867]: INFO nova.compute.manager [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Detaching volume 5c23a596-074b-4f75-a4bc-86c77ffc5080 [ 982.989682] env[61867]: INFO nova.virt.block_device [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Attempting to driver detach volume 5c23a596-074b-4f75-a4bc-86c77ffc5080 from mountpoint /dev/sdb [ 982.990189] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Volume detach. Driver type: vmdk {{(pid=61867) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 982.990637] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274389', 'volume_id': '5c23a596-074b-4f75-a4bc-86c77ffc5080', 'name': 'volume-5c23a596-074b-4f75-a4bc-86c77ffc5080', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2aa08603-d87f-4734-bdfe-fdd610d54e1f', 'attached_at': '', 'detached_at': '', 'volume_id': '5c23a596-074b-4f75-a4bc-86c77ffc5080', 'serial': '5c23a596-074b-4f75-a4bc-86c77ffc5080'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 982.991720] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc5fcd2-e295-4b07-be77-d2fe80541aa8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.020866] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c6fe394-a1e8-4ffe-833a-99349b9ee119 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.029271] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-783b82c8-1ffe-4414-8c7f-def62ae3b54f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.059198] env[61867]: DEBUG nova.network.neutron [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Port e070a089-e0e7-4b11-9688-26d528b37e6a binding to destination host cpu-1 is already ACTIVE {{(pid=61867) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 983.061145] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6de6d3a-e306-48f2-8083-39ac7dafb06d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.081800] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] The volume has not been displaced from its original location: [datastore2] volume-5c23a596-074b-4f75-a4bc-86c77ffc5080/volume-5c23a596-074b-4f75-a4bc-86c77ffc5080.vmdk. No consolidation needed. {{(pid=61867) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 983.087302] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Reconfiguring VM instance instance-0000004d to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 983.089234] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.292s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.089632] env[61867]: DEBUG nova.compute.manager [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 983.096025] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8905c009-3c9e-4729-9d0d-8d731043948d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.108221] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.080s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.110514] env[61867]: INFO nova.compute.claims [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 983.114249] env[61867]: DEBUG oslo_concurrency.lockutils [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "refresh_cache-976299f8-c6cb-4106-90f2-0a4c85625d8c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.114990] env[61867]: DEBUG nova.compute.manager [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Instance network_info: |[{"id": "92925b9b-00ea-4282-9781-7e60212ae093", "address": "fa:16:3e:dd:2b:fa", "network": {"id": "14cee476-296a-4093-99ba-c2a5ef63f7cf", "bridge": "br-int", "label": "tempest-ServersTestJSON-243183479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7316e4f263a9432ab2f9f91484d62f58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92925b9b-00", "ovs_interfaceid": "92925b9b-00ea-4282-9781-7e60212ae093", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 983.114990] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:2b:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '085fb0ff-9285-4f1d-a008-a14da4844357', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '92925b9b-00ea-4282-9781-7e60212ae093', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 983.124643] env[61867]: DEBUG oslo.service.loopingcall [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 983.125883] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 983.126173] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c984ab38-4193-4b34-8d26-1cb29cda3ace {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.144819] env[61867]: DEBUG oslo_vmware.api [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 983.144819] env[61867]: value = "task-1277013" [ 983.144819] env[61867]: _type = "Task" [ 983.144819] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.154636] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 983.154636] env[61867]: value = "task-1277014" [ 983.154636] env[61867]: _type = "Task" [ 983.154636] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.157982] env[61867]: DEBUG nova.compute.manager [req-4a698544-61b1-4363-8b85-920f79690f50 req-7fdb0bf1-b60b-4058-8446-6dc7eb28bd4c service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Received event network-vif-deleted-4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.158311] env[61867]: INFO nova.compute.manager [req-4a698544-61b1-4363-8b85-920f79690f50 req-7fdb0bf1-b60b-4058-8446-6dc7eb28bd4c service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Neutron deleted interface 4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1; detaching it from the instance and deleting it from the info cache [ 983.158421] env[61867]: DEBUG nova.network.neutron [req-4a698544-61b1-4363-8b85-920f79690f50 req-7fdb0bf1-b60b-4058-8446-6dc7eb28bd4c service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.167477] env[61867]: DEBUG oslo_vmware.api [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277013, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.174800] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277014, 'name': CreateVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.303539] env[61867]: DEBUG nova.compute.manager [req-ee4fc98f-7982-4e75-bda6-1e1e423969e9 req-ac501f3b-d1ab-45a6-b686-a6213bf01651 service nova] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Received event network-changed-92925b9b-00ea-4282-9781-7e60212ae093 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.303785] env[61867]: DEBUG nova.compute.manager [req-ee4fc98f-7982-4e75-bda6-1e1e423969e9 req-ac501f3b-d1ab-45a6-b686-a6213bf01651 service nova] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Refreshing instance network info cache due to event network-changed-92925b9b-00ea-4282-9781-7e60212ae093. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 983.304118] env[61867]: DEBUG oslo_concurrency.lockutils [req-ee4fc98f-7982-4e75-bda6-1e1e423969e9 req-ac501f3b-d1ab-45a6-b686-a6213bf01651 service nova] Acquiring lock "refresh_cache-976299f8-c6cb-4106-90f2-0a4c85625d8c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.304369] env[61867]: DEBUG oslo_concurrency.lockutils [req-ee4fc98f-7982-4e75-bda6-1e1e423969e9 req-ac501f3b-d1ab-45a6-b686-a6213bf01651 service nova] Acquired lock "refresh_cache-976299f8-c6cb-4106-90f2-0a4c85625d8c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.304369] env[61867]: DEBUG nova.network.neutron [req-ee4fc98f-7982-4e75-bda6-1e1e423969e9 req-ac501f3b-d1ab-45a6-b686-a6213bf01651 service nova] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Refreshing network info cache for port 92925b9b-00ea-4282-9781-7e60212ae093 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 983.329389] env[61867]: DEBUG oslo_vmware.api [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1277012, 'name': PowerOnVM_Task, 'duration_secs': 1.333301} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.329627] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 983.330250] env[61867]: INFO nova.compute.manager [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Took 8.90 seconds to spawn the instance on the hypervisor. [ 983.330250] env[61867]: DEBUG nova.compute.manager [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 983.330843] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd6b1255-e912-4442-857c-dacc8943ce34 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.343703] env[61867]: INFO nova.compute.manager [-] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Took 1.51 seconds to deallocate network for instance. [ 983.569876] env[61867]: DEBUG nova.network.neutron [-] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.595217] env[61867]: DEBUG nova.compute.utils [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 983.596766] env[61867]: DEBUG nova.compute.manager [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 983.596936] env[61867]: DEBUG nova.network.neutron [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 983.636377] env[61867]: DEBUG nova.policy [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a013ca682054d16a6702230a22c180f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c28a8e133444440eb83dc6a848ef591d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 983.654462] env[61867]: DEBUG oslo_vmware.api [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277013, 'name': ReconfigVM_Task, 'duration_secs': 0.268732} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.654814] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Reconfigured VM instance instance-0000004d to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 983.659786] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-731e4278-1a27-4858-92ca-1badacd78f18 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.673194] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2dcc943f-b83f-4889-961a-df814cf63696 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.681817] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277014, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.683884] env[61867]: DEBUG oslo_vmware.api [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 983.683884] env[61867]: value = "task-1277015" [ 983.683884] env[61867]: _type = "Task" [ 983.683884] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.686628] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c77902e4-67c6-44ac-a6c0-b339b5edcf25 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.704993] env[61867]: DEBUG oslo_vmware.api [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277015, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.716932] env[61867]: DEBUG nova.compute.manager [req-4a698544-61b1-4363-8b85-920f79690f50 req-7fdb0bf1-b60b-4058-8446-6dc7eb28bd4c service nova] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Detach interface failed, port_id=4f63bd0e-c1e2-43b6-a9b6-13083d2ecac1, reason: Instance 7479bf91-5aef-4e75-a127-7e82ae15a003 could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 983.848908] env[61867]: DEBUG oslo_concurrency.lockutils [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.851128] env[61867]: INFO nova.compute.manager [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Took 17.94 seconds to build instance. [ 983.908267] env[61867]: DEBUG nova.network.neutron [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Successfully created port: c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 984.029133] env[61867]: DEBUG nova.network.neutron [req-ee4fc98f-7982-4e75-bda6-1e1e423969e9 req-ac501f3b-d1ab-45a6-b686-a6213bf01651 service nova] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Updated VIF entry in instance network info cache for port 92925b9b-00ea-4282-9781-7e60212ae093. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 984.029522] env[61867]: DEBUG nova.network.neutron [req-ee4fc98f-7982-4e75-bda6-1e1e423969e9 req-ac501f3b-d1ab-45a6-b686-a6213bf01651 service nova] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Updating instance_info_cache with network_info: [{"id": "92925b9b-00ea-4282-9781-7e60212ae093", "address": "fa:16:3e:dd:2b:fa", "network": {"id": "14cee476-296a-4093-99ba-c2a5ef63f7cf", "bridge": "br-int", "label": "tempest-ServersTestJSON-243183479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7316e4f263a9432ab2f9f91484d62f58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92925b9b-00", "ovs_interfaceid": "92925b9b-00ea-4282-9781-7e60212ae093", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.079427] env[61867]: INFO nova.compute.manager [-] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Took 1.78 seconds to deallocate network for instance. [ 984.089204] env[61867]: DEBUG oslo_concurrency.lockutils [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.089204] env[61867]: DEBUG oslo_concurrency.lockutils [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.089204] env[61867]: DEBUG oslo_concurrency.lockutils [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.099411] env[61867]: DEBUG nova.compute.manager [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 984.171599] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277014, 'name': CreateVM_Task, 'duration_secs': 0.639159} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.171775] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 984.172476] env[61867]: DEBUG oslo_concurrency.lockutils [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.172653] env[61867]: DEBUG oslo_concurrency.lockutils [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.173023] env[61867]: DEBUG oslo_concurrency.lockutils [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 984.173244] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6707cbf-41ad-4220-a6b6-b2b474451e28 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.177876] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 984.177876] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d33596-b3ba-8291-4187-af7c7a59b40a" [ 984.177876] env[61867]: _type = "Task" [ 984.177876] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.187829] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d33596-b3ba-8291-4187-af7c7a59b40a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.205851] env[61867]: DEBUG oslo_vmware.api [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277015, 'name': ReconfigVM_Task, 'duration_secs': 0.138011} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.210224] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274389', 'volume_id': '5c23a596-074b-4f75-a4bc-86c77ffc5080', 'name': 'volume-5c23a596-074b-4f75-a4bc-86c77ffc5080', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2aa08603-d87f-4734-bdfe-fdd610d54e1f', 'attached_at': '', 'detached_at': '', 'volume_id': '5c23a596-074b-4f75-a4bc-86c77ffc5080', 'serial': '5c23a596-074b-4f75-a4bc-86c77ffc5080'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 984.304596] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc6fe9f-2663-46e6-a0b2-eadca7f4d1e1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.312300] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dea2bd7-390b-4d55-ab3c-f84935d30f03 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.341451] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6b1575-52a0-4b7d-a590-10c4c09214d6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.348564] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c59def47-9243-4ff4-8a68-e46689c1aecb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.352838] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d558735b-aa83-4745-b209-299153da2b93 tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Lock "d19e7706-e6ac-47dd-8ec3-c30bd315559d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.450s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.364022] env[61867]: DEBUG nova.compute.provider_tree [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.531960] env[61867]: DEBUG oslo_concurrency.lockutils [req-ee4fc98f-7982-4e75-bda6-1e1e423969e9 req-ac501f3b-d1ab-45a6-b686-a6213bf01651 service nova] Releasing lock "refresh_cache-976299f8-c6cb-4106-90f2-0a4c85625d8c" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.532276] env[61867]: DEBUG nova.compute.manager [req-ee4fc98f-7982-4e75-bda6-1e1e423969e9 req-ac501f3b-d1ab-45a6-b686-a6213bf01651 service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Received event network-vif-deleted-86895cfd-1390-4fcd-921e-e39e1ac94da0 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.532520] env[61867]: DEBUG nova.compute.manager [req-ee4fc98f-7982-4e75-bda6-1e1e423969e9 req-ac501f3b-d1ab-45a6-b686-a6213bf01651 service nova] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Received event network-vif-deleted-bb767aa8-7bfe-4604-9152-059b0eaa3311 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.592899] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.687832] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d33596-b3ba-8291-4187-af7c7a59b40a, 'name': SearchDatastore_Task, 'duration_secs': 0.009065} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.688142] env[61867]: DEBUG oslo_concurrency.lockutils [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.688374] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 984.688613] env[61867]: DEBUG oslo_concurrency.lockutils [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.688763] env[61867]: DEBUG oslo_concurrency.lockutils [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.689040] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 984.689215] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e1de0322-55ea-4732-b266-62ac1efa817f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.697366] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 984.697549] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 984.698225] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a699e095-a4d9-48b8-beef-f11b0329f390 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.703508] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 984.703508] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526027c3-382f-b8a4-2379-4b7a5a2e7a21" [ 984.703508] env[61867]: _type = "Task" [ 984.703508] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.710597] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526027c3-382f-b8a4-2379-4b7a5a2e7a21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.763908] env[61867]: DEBUG nova.objects.instance [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lazy-loading 'flavor' on Instance uuid 2aa08603-d87f-4734-bdfe-fdd610d54e1f {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.866416] env[61867]: DEBUG nova.scheduler.client.report [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 985.107202] env[61867]: DEBUG nova.compute.manager [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 985.110112] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5468cfb1-9486-4b4c-9226-98d07aed6fee tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Acquiring lock "interface-d19e7706-e6ac-47dd-8ec3-c30bd315559d-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.110393] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5468cfb1-9486-4b4c-9226-98d07aed6fee tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Lock "interface-d19e7706-e6ac-47dd-8ec3-c30bd315559d-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.110805] env[61867]: DEBUG nova.objects.instance [None req-5468cfb1-9486-4b4c-9226-98d07aed6fee tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Lazy-loading 'flavor' on Instance uuid d19e7706-e6ac-47dd-8ec3-c30bd315559d {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.124253] env[61867]: DEBUG oslo_concurrency.lockutils [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "refresh_cache-6e2136e7-c7e0-4a98-9899-f79c10f0e703" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.124468] env[61867]: DEBUG oslo_concurrency.lockutils [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "refresh_cache-6e2136e7-c7e0-4a98-9899-f79c10f0e703" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.124687] env[61867]: DEBUG nova.network.neutron [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 985.135387] env[61867]: DEBUG nova.virt.hardware [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 985.135712] env[61867]: DEBUG nova.virt.hardware [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 985.135952] env[61867]: DEBUG nova.virt.hardware [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.136429] env[61867]: DEBUG nova.virt.hardware [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 985.136429] env[61867]: DEBUG nova.virt.hardware [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.136547] env[61867]: DEBUG nova.virt.hardware [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 985.136746] env[61867]: DEBUG nova.virt.hardware [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 985.136959] env[61867]: DEBUG nova.virt.hardware [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 985.137215] env[61867]: DEBUG nova.virt.hardware [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 985.137397] env[61867]: DEBUG nova.virt.hardware [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 985.137577] env[61867]: DEBUG nova.virt.hardware [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.138697] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7090244-1ec6-4ba4-826e-035aa7305a9a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.147945] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e38f07-911b-43cd-9ab8-eaf1e3b704e9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.214302] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526027c3-382f-b8a4-2379-4b7a5a2e7a21, 'name': SearchDatastore_Task, 'duration_secs': 0.008064} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.215077] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-499a1d21-70b5-49dc-9be0-0450bfcf0910 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.220471] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 985.220471] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520dfa81-4bbe-42cb-b8a0-74ac755a719d" [ 985.220471] env[61867]: _type = "Task" [ 985.220471] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.228251] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520dfa81-4bbe-42cb-b8a0-74ac755a719d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.373075] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.264s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.373246] env[61867]: DEBUG nova.compute.manager [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 985.377891] env[61867]: DEBUG oslo_concurrency.lockutils [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.529s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.378135] env[61867]: DEBUG nova.objects.instance [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lazy-loading 'resources' on Instance uuid f426dd3e-2806-4d70-8784-75c611fd03e4 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.523133] env[61867]: DEBUG nova.compute.manager [req-f9c459a6-d741-440c-821b-267176a8d5f4 req-ced86149-782a-4a24-85e7-92b82461011a service nova] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Received event network-vif-plugged-c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.523395] env[61867]: DEBUG oslo_concurrency.lockutils [req-f9c459a6-d741-440c-821b-267176a8d5f4 req-ced86149-782a-4a24-85e7-92b82461011a service nova] Acquiring lock "4006cabd-cb6e-45b6-b87e-9b4b1550f4fe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.523616] env[61867]: DEBUG oslo_concurrency.lockutils [req-f9c459a6-d741-440c-821b-267176a8d5f4 req-ced86149-782a-4a24-85e7-92b82461011a service nova] Lock "4006cabd-cb6e-45b6-b87e-9b4b1550f4fe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.523792] env[61867]: DEBUG oslo_concurrency.lockutils [req-f9c459a6-d741-440c-821b-267176a8d5f4 req-ced86149-782a-4a24-85e7-92b82461011a service nova] Lock "4006cabd-cb6e-45b6-b87e-9b4b1550f4fe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.523965] env[61867]: DEBUG nova.compute.manager [req-f9c459a6-d741-440c-821b-267176a8d5f4 req-ced86149-782a-4a24-85e7-92b82461011a service nova] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] No waiting events found dispatching network-vif-plugged-c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 985.524462] env[61867]: WARNING nova.compute.manager [req-f9c459a6-d741-440c-821b-267176a8d5f4 req-ced86149-782a-4a24-85e7-92b82461011a service nova] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Received unexpected event network-vif-plugged-c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78 for instance with vm_state building and task_state spawning. [ 985.614734] env[61867]: DEBUG nova.objects.instance [None req-5468cfb1-9486-4b4c-9226-98d07aed6fee tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Lazy-loading 'pci_requests' on Instance uuid d19e7706-e6ac-47dd-8ec3-c30bd315559d {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.616443] env[61867]: DEBUG nova.network.neutron [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Successfully updated port: c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 985.731691] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520dfa81-4bbe-42cb-b8a0-74ac755a719d, 'name': SearchDatastore_Task, 'duration_secs': 0.009803} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.731997] env[61867]: DEBUG oslo_concurrency.lockutils [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.732286] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 976299f8-c6cb-4106-90f2-0a4c85625d8c/976299f8-c6cb-4106-90f2-0a4c85625d8c.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 985.732633] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6769e955-4fae-4aa9-b575-7790f54c8a70 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.739796] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 985.739796] env[61867]: value = "task-1277016" [ 985.739796] env[61867]: _type = "Task" [ 985.739796] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.750425] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1277016, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.771638] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1ff3af2f-9c31-43fb-8279-bd9a0ac6acb0 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.323s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.845453] env[61867]: DEBUG nova.network.neutron [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Updating instance_info_cache with network_info: [{"id": "e070a089-e0e7-4b11-9688-26d528b37e6a", "address": "fa:16:3e:7b:ac:a0", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape070a089-e0", "ovs_interfaceid": "e070a089-e0e7-4b11-9688-26d528b37e6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.881654] env[61867]: DEBUG nova.compute.utils [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 985.883150] env[61867]: DEBUG nova.compute.manager [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 985.883325] env[61867]: DEBUG nova.network.neutron [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 985.958465] env[61867]: DEBUG nova.policy [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1fc8ff85bced44c3a9de8b6c71d02bd5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '250e5d2fa5cb4a8a8bd0a6fdf9e50a3d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 986.119203] env[61867]: DEBUG nova.objects.base [None req-5468cfb1-9486-4b4c-9226-98d07aed6fee tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61867) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 986.119512] env[61867]: DEBUG nova.network.neutron [None req-5468cfb1-9486-4b4c-9226-98d07aed6fee tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 986.122390] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "refresh_cache-4006cabd-cb6e-45b6-b87e-9b4b1550f4fe" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.122663] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "refresh_cache-4006cabd-cb6e-45b6-b87e-9b4b1550f4fe" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.123112] env[61867]: DEBUG nova.network.neutron [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.195799] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02737ef0-fe88-417e-bb45-fdc437686e71 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.202987] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76783b3f-2761-4775-b474-a3cb21aaa3ff {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.209562] env[61867]: DEBUG oslo_concurrency.lockutils [None req-5468cfb1-9486-4b4c-9226-98d07aed6fee tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Lock "interface-d19e7706-e6ac-47dd-8ec3-c30bd315559d-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.099s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.237079] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0331500-69e2-48ef-a29e-8c7a83512bcf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.251260] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c64a47-b0f0-4f35-8a3a-bb2d0461f306 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.258213] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1277016, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.435241} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.265869] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 976299f8-c6cb-4106-90f2-0a4c85625d8c/976299f8-c6cb-4106-90f2-0a4c85625d8c.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 986.266142] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 986.266642] env[61867]: DEBUG nova.compute.provider_tree [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.268353] env[61867]: DEBUG nova.network.neutron [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Successfully created port: f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 986.270117] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8a70b663-94d5-41ee-87a6-d39cfd7fc39f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.277549] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 986.277549] env[61867]: value = "task-1277017" [ 986.277549] env[61867]: _type = "Task" [ 986.277549] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.286407] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1277017, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.348904] env[61867]: DEBUG oslo_concurrency.lockutils [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "refresh_cache-6e2136e7-c7e0-4a98-9899-f79c10f0e703" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.388827] env[61867]: DEBUG nova.compute.manager [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 986.458902] env[61867]: DEBUG oslo_concurrency.lockutils [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "interface-4d7bfb31-d565-49e4-8c51-0122acebba2d-aa5b1005-fdcb-4208-b0d8-3b33573d05b6" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.458902] env[61867]: DEBUG oslo_concurrency.lockutils [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "interface-4d7bfb31-d565-49e4-8c51-0122acebba2d-aa5b1005-fdcb-4208-b0d8-3b33573d05b6" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.459116] env[61867]: DEBUG nova.objects.instance [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lazy-loading 'flavor' on Instance uuid 4d7bfb31-d565-49e4-8c51-0122acebba2d {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.657668] env[61867]: DEBUG nova.network.neutron [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 986.773695] env[61867]: DEBUG nova.scheduler.client.report [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.784915] env[61867]: DEBUG nova.network.neutron [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Updating instance_info_cache with network_info: [{"id": "c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78", "address": "fa:16:3e:70:a2:ef", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5ee9c92-b2", "ovs_interfaceid": "c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.789764] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1277017, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062912} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.790304] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 986.791397] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e47c4855-c9fc-4eb7-a21b-abf6f49c20fc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.813744] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 976299f8-c6cb-4106-90f2-0a4c85625d8c/976299f8-c6cb-4106-90f2-0a4c85625d8c.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 986.814154] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5521fa2-b567-4fdd-802e-8b6573a2f8d7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.836390] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 986.836390] env[61867]: value = "task-1277018" [ 986.836390] env[61867]: _type = "Task" [ 986.836390] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.844502] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1277018, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.872525] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e85ae78c-1fc9-493e-afef-90a5f400245d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.892270] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8ab561-f3f8-4e5b-9cac-c5b578657821 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.903079] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Updating instance '6e2136e7-c7e0-4a98-9899-f79c10f0e703' progress to 83 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 987.054521] env[61867]: DEBUG nova.objects.instance [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lazy-loading 'pci_requests' on Instance uuid 4d7bfb31-d565-49e4-8c51-0122acebba2d {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.278539] env[61867]: DEBUG oslo_concurrency.lockutils [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.900s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.281143] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.688s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.281390] env[61867]: DEBUG nova.objects.instance [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lazy-loading 'resources' on Instance uuid 7479bf91-5aef-4e75-a127-7e82ae15a003 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.291240] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "refresh_cache-4006cabd-cb6e-45b6-b87e-9b4b1550f4fe" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.291533] env[61867]: DEBUG nova.compute.manager [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Instance network_info: |[{"id": "c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78", "address": "fa:16:3e:70:a2:ef", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5ee9c92-b2", "ovs_interfaceid": "c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 987.292190] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:a2:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74e6f6e0-95e6-4531-99e9-0e78350fb655', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.299712] env[61867]: DEBUG oslo.service.loopingcall [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.300743] env[61867]: INFO nova.scheduler.client.report [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Deleted allocations for instance f426dd3e-2806-4d70-8784-75c611fd03e4 [ 987.301846] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 987.304200] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1b76670e-561a-476b-8eb2-6a8a61eac219 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.324836] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.324836] env[61867]: value = "task-1277019" [ 987.324836] env[61867]: _type = "Task" [ 987.324836] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.333455] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277019, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.346357] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1277018, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.400880] env[61867]: DEBUG nova.compute.manager [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 987.408943] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 987.409423] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e6a69f4d-e5be-4bf8-aa09-0df76fd3ea65 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.417938] env[61867]: DEBUG oslo_vmware.api [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 987.417938] env[61867]: value = "task-1277020" [ 987.417938] env[61867]: _type = "Task" [ 987.417938] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.426130] env[61867]: DEBUG nova.virt.hardware [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.426384] env[61867]: DEBUG nova.virt.hardware [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.426552] env[61867]: DEBUG nova.virt.hardware [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.426746] env[61867]: DEBUG nova.virt.hardware [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.426898] env[61867]: DEBUG nova.virt.hardware [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.427897] env[61867]: DEBUG nova.virt.hardware [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.427897] env[61867]: DEBUG nova.virt.hardware [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.427897] env[61867]: DEBUG nova.virt.hardware [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.427897] env[61867]: DEBUG nova.virt.hardware [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.427897] env[61867]: DEBUG nova.virt.hardware [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.427897] env[61867]: DEBUG nova.virt.hardware [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.428988] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2374b4e0-0850-431e-b3f5-c0255a8358d8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.433957] env[61867]: DEBUG oslo_vmware.api [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277020, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.438949] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d842e87-5ad5-4338-8380-0e490f8b92a5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.557416] env[61867]: DEBUG nova.objects.base [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Object Instance<4d7bfb31-d565-49e4-8c51-0122acebba2d> lazy-loaded attributes: flavor,pci_requests {{(pid=61867) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 987.557711] env[61867]: DEBUG nova.network.neutron [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 987.633106] env[61867]: DEBUG nova.policy [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ea634246436422a9407f82692b69ede', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '86ced20eaf4740e298dc6f8ca5550c09', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 987.824182] env[61867]: DEBUG oslo_concurrency.lockutils [None req-40f13be2-0783-46e3-9506-87b895fc8f03 tempest-ServersTestMultiNic-351719036 tempest-ServersTestMultiNic-351719036-project-member] Lock "f426dd3e-2806-4d70-8784-75c611fd03e4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.177s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.832584] env[61867]: DEBUG nova.network.neutron [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Successfully updated port: f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 987.840510] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277019, 'name': CreateVM_Task, 'duration_secs': 0.389566} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.843501] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 987.845721] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.845907] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.846508] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 987.847537] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65100261-9b5f-46e3-a7f6-519f77bbd03e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.853852] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1277018, 'name': ReconfigVM_Task, 'duration_secs': 0.582012} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.857017] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 976299f8-c6cb-4106-90f2-0a4c85625d8c/976299f8-c6cb-4106-90f2-0a4c85625d8c.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.858092] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b8d3ed08-b19e-4ffe-9242-ff3d932da5de {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.860716] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 987.860716] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52eee9d7-a0e3-fa2b-0283-ebf50803df71" [ 987.860716] env[61867]: _type = "Task" [ 987.860716] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.865189] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 987.865189] env[61867]: value = "task-1277021" [ 987.865189] env[61867]: _type = "Task" [ 987.865189] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.871201] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52eee9d7-a0e3-fa2b-0283-ebf50803df71, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.878529] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1277021, 'name': Rename_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.928962] env[61867]: DEBUG oslo_vmware.api [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277020, 'name': PowerOnVM_Task, 'duration_secs': 0.376493} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.931411] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 987.931622] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-97f8fb48-9910-4fe3-9706-e609a7cc6861 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Updating instance '6e2136e7-c7e0-4a98-9899-f79c10f0e703' progress to 100 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 987.988713] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be5af691-eff1-4926-9e13-a5e74231066a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.997062] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd532c6-5361-48b6-9d6e-745cef0c64b5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.026852] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72b3b23-7786-48e5-8c91-44c71ffa8e77 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.035057] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2671a696-71ba-43de-a168-f2e9b1067140 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.048620] env[61867]: DEBUG nova.compute.provider_tree [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.255015] env[61867]: DEBUG nova.compute.manager [req-cb4847af-c124-4266-af4a-3d82632f186a req-da9ea1a7-ef46-4210-ac3c-6848cd916a3e service nova] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Received event network-changed-c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.255238] env[61867]: DEBUG nova.compute.manager [req-cb4847af-c124-4266-af4a-3d82632f186a req-da9ea1a7-ef46-4210-ac3c-6848cd916a3e service nova] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Refreshing instance network info cache due to event network-changed-c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 988.255459] env[61867]: DEBUG oslo_concurrency.lockutils [req-cb4847af-c124-4266-af4a-3d82632f186a req-da9ea1a7-ef46-4210-ac3c-6848cd916a3e service nova] Acquiring lock "refresh_cache-4006cabd-cb6e-45b6-b87e-9b4b1550f4fe" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.255605] env[61867]: DEBUG oslo_concurrency.lockutils [req-cb4847af-c124-4266-af4a-3d82632f186a req-da9ea1a7-ef46-4210-ac3c-6848cd916a3e service nova] Acquired lock "refresh_cache-4006cabd-cb6e-45b6-b87e-9b4b1550f4fe" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.255765] env[61867]: DEBUG nova.network.neutron [req-cb4847af-c124-4266-af4a-3d82632f186a req-da9ea1a7-ef46-4210-ac3c-6848cd916a3e service nova] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Refreshing network info cache for port c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 988.335631] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.335631] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.335631] env[61867]: DEBUG nova.network.neutron [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 988.375761] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1277021, 'name': Rename_Task, 'duration_secs': 0.141397} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.381766] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 988.382341] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52eee9d7-a0e3-fa2b-0283-ebf50803df71, 'name': SearchDatastore_Task, 'duration_secs': 0.011943} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.382726] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3ad6602a-4fed-4639-b3b4-d2c8ca77075f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.384362] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.384595] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 988.384827] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.385043] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.385276] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 988.386041] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23a5b840-daa5-4119-a9da-95d19a28ab91 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.392737] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 988.392737] env[61867]: value = "task-1277022" [ 988.392737] env[61867]: _type = "Task" [ 988.392737] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.398154] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 988.398347] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 988.399910] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99e2f36b-53a1-46e1-8b57-4d0a7cf7cca5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.405676] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1277022, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.408462] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 988.408462] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e71619-9fed-f471-4eb5-dc214e4e91ee" [ 988.408462] env[61867]: _type = "Task" [ 988.408462] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.416185] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e71619-9fed-f471-4eb5-dc214e4e91ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.551851] env[61867]: DEBUG nova.scheduler.client.report [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 988.812439] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Acquiring lock "d19e7706-e6ac-47dd-8ec3-c30bd315559d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.812439] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Lock "d19e7706-e6ac-47dd-8ec3-c30bd315559d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.812439] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Acquiring lock "d19e7706-e6ac-47dd-8ec3-c30bd315559d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.812439] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Lock "d19e7706-e6ac-47dd-8ec3-c30bd315559d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.812439] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Lock "d19e7706-e6ac-47dd-8ec3-c30bd315559d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.813738] env[61867]: INFO nova.compute.manager [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Terminating instance [ 988.815679] env[61867]: DEBUG nova.compute.manager [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 988.819021] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 988.819021] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c16fc92-5537-4fa7-929f-7d20ec8f51b1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.825869] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 988.826409] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5316b815-9662-4da5-bd79-30971e68eec6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.835122] env[61867]: DEBUG oslo_vmware.api [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Waiting for the task: (returnval){ [ 988.835122] env[61867]: value = "task-1277023" [ 988.835122] env[61867]: _type = "Task" [ 988.835122] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.847380] env[61867]: DEBUG oslo_vmware.api [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1277023, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.880317] env[61867]: DEBUG nova.network.neutron [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 988.884284] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.884284] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.884443] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.885019] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.885019] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.886789] env[61867]: INFO nova.compute.manager [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Terminating instance [ 988.888573] env[61867]: DEBUG nova.compute.manager [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 988.888713] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 988.889539] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad82db5-cd76-4987-a2b4-740f0ab15a1c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.903576] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 988.907880] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-df708fee-0327-45a9-8851-075ccbae50df {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.920470] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1277022, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.922574] env[61867]: DEBUG oslo_vmware.api [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 988.922574] env[61867]: value = "task-1277024" [ 988.922574] env[61867]: _type = "Task" [ 988.922574] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.929891] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e71619-9fed-f471-4eb5-dc214e4e91ee, 'name': SearchDatastore_Task, 'duration_secs': 0.012547} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.931198] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6b028b8-30f0-4ebc-a9c1-4485eb74656e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.936891] env[61867]: DEBUG oslo_vmware.api [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277024, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.944535] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 988.944535] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52a96dd7-eb18-45da-afa7-d37f18e92fd3" [ 988.944535] env[61867]: _type = "Task" [ 988.944535] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.959741] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52a96dd7-eb18-45da-afa7-d37f18e92fd3, 'name': SearchDatastore_Task, 'duration_secs': 0.010954} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.963020] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.963020] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe/4006cabd-cb6e-45b6-b87e-9b4b1550f4fe.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 988.963020] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-37e24f03-4a51-4332-bbf0-f3b61a6cde10 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.966909] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 988.966909] env[61867]: value = "task-1277025" [ 988.966909] env[61867]: _type = "Task" [ 988.966909] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.976010] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277025, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.057434] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.776s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.063340] env[61867]: DEBUG nova.network.neutron [req-cb4847af-c124-4266-af4a-3d82632f186a req-da9ea1a7-ef46-4210-ac3c-6848cd916a3e service nova] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Updated VIF entry in instance network info cache for port c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 989.063816] env[61867]: DEBUG nova.network.neutron [req-cb4847af-c124-4266-af4a-3d82632f186a req-da9ea1a7-ef46-4210-ac3c-6848cd916a3e service nova] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Updating instance_info_cache with network_info: [{"id": "c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78", "address": "fa:16:3e:70:a2:ef", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5ee9c92-b2", "ovs_interfaceid": "c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.091651] env[61867]: INFO nova.scheduler.client.report [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Deleted allocations for instance 7479bf91-5aef-4e75-a127-7e82ae15a003 [ 989.127154] env[61867]: DEBUG nova.network.neutron [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating instance_info_cache with network_info: [{"id": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "address": "fa:16:3e:b9:62:3b", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1e84fbf-9f", "ovs_interfaceid": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.342856] env[61867]: DEBUG nova.compute.manager [req-6392dc1d-06ca-45e5-9ce2-87a291983d82 req-61a7ef4a-932c-44e3-8406-53651ffaba26 service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Received event network-vif-plugged-aa5b1005-fdcb-4208-b0d8-3b33573d05b6 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.343129] env[61867]: DEBUG oslo_concurrency.lockutils [req-6392dc1d-06ca-45e5-9ce2-87a291983d82 req-61a7ef4a-932c-44e3-8406-53651ffaba26 service nova] Acquiring lock "4d7bfb31-d565-49e4-8c51-0122acebba2d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.343690] env[61867]: DEBUG oslo_concurrency.lockutils [req-6392dc1d-06ca-45e5-9ce2-87a291983d82 req-61a7ef4a-932c-44e3-8406-53651ffaba26 service nova] Lock "4d7bfb31-d565-49e4-8c51-0122acebba2d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.343890] env[61867]: DEBUG oslo_concurrency.lockutils [req-6392dc1d-06ca-45e5-9ce2-87a291983d82 req-61a7ef4a-932c-44e3-8406-53651ffaba26 service nova] Lock "4d7bfb31-d565-49e4-8c51-0122acebba2d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.344080] env[61867]: DEBUG nova.compute.manager [req-6392dc1d-06ca-45e5-9ce2-87a291983d82 req-61a7ef4a-932c-44e3-8406-53651ffaba26 service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] No waiting events found dispatching network-vif-plugged-aa5b1005-fdcb-4208-b0d8-3b33573d05b6 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 989.344255] env[61867]: WARNING nova.compute.manager [req-6392dc1d-06ca-45e5-9ce2-87a291983d82 req-61a7ef4a-932c-44e3-8406-53651ffaba26 service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Received unexpected event network-vif-plugged-aa5b1005-fdcb-4208-b0d8-3b33573d05b6 for instance with vm_state active and task_state None. [ 989.351045] env[61867]: DEBUG oslo_vmware.api [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1277023, 'name': PowerOffVM_Task, 'duration_secs': 0.209408} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.351346] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 989.351485] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 989.352076] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d6b9189-44f9-4fe6-85ba-d9d4e1ae482f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.354306] env[61867]: DEBUG nova.network.neutron [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Successfully updated port: aa5b1005-fdcb-4208-b0d8-3b33573d05b6 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 989.407948] env[61867]: DEBUG oslo_vmware.api [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1277022, 'name': PowerOnVM_Task, 'duration_secs': 0.811651} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.408686] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 989.409231] env[61867]: INFO nova.compute.manager [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Took 8.42 seconds to spawn the instance on the hypervisor. [ 989.409647] env[61867]: DEBUG nova.compute.manager [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 989.410786] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cceee7c0-422f-43b4-aea7-15b2cd3686fd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.432593] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 989.432859] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 989.433062] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Deleting the datastore file [datastore1] d19e7706-e6ac-47dd-8ec3-c30bd315559d {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 989.436452] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-888a0fd3-b011-473b-9e08-c07d979b658f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.438495] env[61867]: DEBUG oslo_vmware.api [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277024, 'name': PowerOffVM_Task, 'duration_secs': 0.198822} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.438849] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 989.438949] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 989.439629] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eafd571b-cb00-4a4d-bc9b-5ec0dc0be862 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.443528] env[61867]: DEBUG oslo_vmware.api [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Waiting for the task: (returnval){ [ 989.443528] env[61867]: value = "task-1277027" [ 989.443528] env[61867]: _type = "Task" [ 989.443528] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.461540] env[61867]: DEBUG oslo_vmware.api [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1277027, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.478399] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277025, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.501879] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 989.502124] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 989.502315] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Deleting the datastore file [datastore2] 2aa08603-d87f-4734-bdfe-fdd610d54e1f {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 989.502638] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-16beba2c-e092-414f-b677-708683f221a6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.508631] env[61867]: DEBUG oslo_vmware.api [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 989.508631] env[61867]: value = "task-1277029" [ 989.508631] env[61867]: _type = "Task" [ 989.508631] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.516378] env[61867]: DEBUG oslo_vmware.api [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277029, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.571578] env[61867]: DEBUG oslo_concurrency.lockutils [req-cb4847af-c124-4266-af4a-3d82632f186a req-da9ea1a7-ef46-4210-ac3c-6848cd916a3e service nova] Releasing lock "refresh_cache-4006cabd-cb6e-45b6-b87e-9b4b1550f4fe" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.600255] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fba42d3-6bc1-4940-a6f7-f6be40d3e607 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "7479bf91-5aef-4e75-a127-7e82ae15a003" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.427s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.630034] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.630600] env[61867]: DEBUG nova.compute.manager [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Instance network_info: |[{"id": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "address": "fa:16:3e:b9:62:3b", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1e84fbf-9f", "ovs_interfaceid": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 989.630806] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:62:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '399f3826-705c-45f7-9fe0-3a08a945151a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 989.638283] env[61867]: DEBUG oslo.service.loopingcall [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.638489] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 989.638706] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d41e7379-8d3b-44ca-9b5d-262f46d94c8e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.660097] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 989.660097] env[61867]: value = "task-1277030" [ 989.660097] env[61867]: _type = "Task" [ 989.660097] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.670387] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277030, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.864378] env[61867]: DEBUG oslo_concurrency.lockutils [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.864643] env[61867]: DEBUG oslo_concurrency.lockutils [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.864897] env[61867]: DEBUG nova.network.neutron [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 989.935979] env[61867]: INFO nova.compute.manager [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Took 18.04 seconds to build instance. [ 989.954750] env[61867]: DEBUG oslo_vmware.api [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1277027, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.981168] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277025, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.573065} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.981168] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe/4006cabd-cb6e-45b6-b87e-9b4b1550f4fe.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 989.981168] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 989.981168] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f743fcd-0aa4-4a26-b97e-5409430f8349 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.987862] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 989.987862] env[61867]: value = "task-1277031" [ 989.987862] env[61867]: _type = "Task" [ 989.987862] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.001281] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277031, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.023258] env[61867]: DEBUG oslo_vmware.api [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277029, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.392607} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.023258] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 990.023258] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 990.023579] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 990.024598] env[61867]: INFO nova.compute.manager [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Took 1.13 seconds to destroy the instance on the hypervisor. [ 990.024598] env[61867]: DEBUG oslo.service.loopingcall [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.024598] env[61867]: DEBUG nova.compute.manager [-] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 990.024598] env[61867]: DEBUG nova.network.neutron [-] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 990.172318] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277030, 'name': CreateVM_Task, 'duration_secs': 0.452152} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.172585] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 990.173384] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.173599] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.173959] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 990.174446] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0e8f6a3-654f-42d2-890a-9212ff48733f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.179545] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 990.179545] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522608fa-9f71-755f-13b3-a9195711d41e" [ 990.179545] env[61867]: _type = "Task" [ 990.179545] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.188015] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522608fa-9f71-755f-13b3-a9195711d41e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.221353] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.221643] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.221840] env[61867]: DEBUG nova.compute.manager [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Going to confirm migration 4 {{(pid=61867) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 990.316234] env[61867]: DEBUG nova.compute.manager [req-5b819e60-4d70-441a-8c60-b8afda5a0a55 req-8008b0c6-432f-481a-a608-5c5cc76c2edb service nova] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Received event network-vif-plugged-f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.316480] env[61867]: DEBUG oslo_concurrency.lockutils [req-5b819e60-4d70-441a-8c60-b8afda5a0a55 req-8008b0c6-432f-481a-a608-5c5cc76c2edb service nova] Acquiring lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.316696] env[61867]: DEBUG oslo_concurrency.lockutils [req-5b819e60-4d70-441a-8c60-b8afda5a0a55 req-8008b0c6-432f-481a-a608-5c5cc76c2edb service nova] Lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.316867] env[61867]: DEBUG oslo_concurrency.lockutils [req-5b819e60-4d70-441a-8c60-b8afda5a0a55 req-8008b0c6-432f-481a-a608-5c5cc76c2edb service nova] Lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.317441] env[61867]: DEBUG nova.compute.manager [req-5b819e60-4d70-441a-8c60-b8afda5a0a55 req-8008b0c6-432f-481a-a608-5c5cc76c2edb service nova] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] No waiting events found dispatching network-vif-plugged-f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 990.317441] env[61867]: WARNING nova.compute.manager [req-5b819e60-4d70-441a-8c60-b8afda5a0a55 req-8008b0c6-432f-481a-a608-5c5cc76c2edb service nova] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Received unexpected event network-vif-plugged-f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2 for instance with vm_state building and task_state spawning. [ 990.317441] env[61867]: DEBUG nova.compute.manager [req-5b819e60-4d70-441a-8c60-b8afda5a0a55 req-8008b0c6-432f-481a-a608-5c5cc76c2edb service nova] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Received event network-changed-f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.317590] env[61867]: DEBUG nova.compute.manager [req-5b819e60-4d70-441a-8c60-b8afda5a0a55 req-8008b0c6-432f-481a-a608-5c5cc76c2edb service nova] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Refreshing instance network info cache due to event network-changed-f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 990.317733] env[61867]: DEBUG oslo_concurrency.lockutils [req-5b819e60-4d70-441a-8c60-b8afda5a0a55 req-8008b0c6-432f-481a-a608-5c5cc76c2edb service nova] Acquiring lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.317874] env[61867]: DEBUG oslo_concurrency.lockutils [req-5b819e60-4d70-441a-8c60-b8afda5a0a55 req-8008b0c6-432f-481a-a608-5c5cc76c2edb service nova] Acquired lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.318044] env[61867]: DEBUG nova.network.neutron [req-5b819e60-4d70-441a-8c60-b8afda5a0a55 req-8008b0c6-432f-481a-a608-5c5cc76c2edb service nova] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Refreshing network info cache for port f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 990.438709] env[61867]: DEBUG oslo_concurrency.lockutils [None req-834240b2-4a86-4357-a0e8-530ea01d38db tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "976299f8-c6cb-4106-90f2-0a4c85625d8c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.556s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.456261] env[61867]: DEBUG oslo_vmware.api [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Task: {'id': task-1277027, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.595696} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.456261] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 990.456261] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 990.456594] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 990.456699] env[61867]: INFO nova.compute.manager [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Took 1.64 seconds to destroy the instance on the hypervisor. [ 990.457009] env[61867]: DEBUG oslo.service.loopingcall [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.457246] env[61867]: DEBUG nova.compute.manager [-] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 990.457880] env[61867]: DEBUG nova.network.neutron [-] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 990.500390] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277031, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086307} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.500810] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 990.501669] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313bd30d-f69a-4047-bf05-3866142589f8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.531309] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe/4006cabd-cb6e-45b6-b87e-9b4b1550f4fe.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 990.531765] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-649078f4-bff8-40bc-8292-f1dd6dbe6297 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.556769] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 990.556769] env[61867]: value = "task-1277033" [ 990.556769] env[61867]: _type = "Task" [ 990.556769] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.567127] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277033, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.630994] env[61867]: WARNING nova.network.neutron [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] 7a62cd67-4e7d-4952-9a24-053e35c1c3cb already exists in list: networks containing: ['7a62cd67-4e7d-4952-9a24-053e35c1c3cb']. ignoring it [ 990.698547] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]522608fa-9f71-755f-13b3-a9195711d41e, 'name': SearchDatastore_Task, 'duration_secs': 0.021835} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.698866] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.699255] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 990.699512] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.699663] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.699845] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 990.700128] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dbbe7120-b969-4023-91d0-3b393e96f0a3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.710485] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 990.710485] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 990.710618] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6c9cfdd-e5c8-4d5b-9be7-89c5c26f3463 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.717019] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 990.717019] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d7a02c-4753-d360-09fa-57e3c06bce8e" [ 990.717019] env[61867]: _type = "Task" [ 990.717019] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.731456] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d7a02c-4753-d360-09fa-57e3c06bce8e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.939158] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "refresh_cache-6e2136e7-c7e0-4a98-9899-f79c10f0e703" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.939444] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "refresh_cache-6e2136e7-c7e0-4a98-9899-f79c10f0e703" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.939588] env[61867]: DEBUG nova.network.neutron [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 990.940065] env[61867]: DEBUG nova.objects.instance [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lazy-loading 'info_cache' on Instance uuid 6e2136e7-c7e0-4a98-9899-f79c10f0e703 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.067767] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277033, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.231822] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d7a02c-4753-d360-09fa-57e3c06bce8e, 'name': SearchDatastore_Task, 'duration_secs': 0.021428} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.232863] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e067d93-760c-4758-84f8-90327b6be7d5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.247234] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 991.247234] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528276a7-e663-bdb4-396f-4abd1975e5e4" [ 991.247234] env[61867]: _type = "Task" [ 991.247234] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.256573] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528276a7-e663-bdb4-396f-4abd1975e5e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.315998] env[61867]: DEBUG nova.network.neutron [req-5b819e60-4d70-441a-8c60-b8afda5a0a55 req-8008b0c6-432f-481a-a608-5c5cc76c2edb service nova] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updated VIF entry in instance network info cache for port f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 991.316398] env[61867]: DEBUG nova.network.neutron [req-5b819e60-4d70-441a-8c60-b8afda5a0a55 req-8008b0c6-432f-481a-a608-5c5cc76c2edb service nova] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating instance_info_cache with network_info: [{"id": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "address": "fa:16:3e:b9:62:3b", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1e84fbf-9f", "ovs_interfaceid": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.330784] env[61867]: DEBUG nova.network.neutron [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Updating instance_info_cache with network_info: [{"id": "c05702c4-5138-450d-82b2-790b86d8a59b", "address": "fa:16:3e:cf:d4:2a", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc05702c4-51", "ovs_interfaceid": "c05702c4-5138-450d-82b2-790b86d8a59b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "aa5b1005-fdcb-4208-b0d8-3b33573d05b6", "address": "fa:16:3e:80:74:6e", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa5b1005-fd", "ovs_interfaceid": "aa5b1005-fdcb-4208-b0d8-3b33573d05b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.423009] env[61867]: DEBUG nova.network.neutron [-] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.567476] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277033, 'name': ReconfigVM_Task, 'duration_secs': 0.603513} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.567765] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe/4006cabd-cb6e-45b6-b87e-9b4b1550f4fe.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 991.568434] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e00ddca6-b8de-4b97-9971-9d9291e7c639 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.574111] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 991.574111] env[61867]: value = "task-1277034" [ 991.574111] env[61867]: _type = "Task" [ 991.574111] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.582019] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277034, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.705441] env[61867]: DEBUG nova.compute.manager [req-41b7dd23-3ad2-46f5-a04f-adf0cc3c033b req-eff04d60-0f66-41af-81d6-ac62440ba2af service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Received event network-changed-aa5b1005-fdcb-4208-b0d8-3b33573d05b6 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.705441] env[61867]: DEBUG nova.compute.manager [req-41b7dd23-3ad2-46f5-a04f-adf0cc3c033b req-eff04d60-0f66-41af-81d6-ac62440ba2af service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Refreshing instance network info cache due to event network-changed-aa5b1005-fdcb-4208-b0d8-3b33573d05b6. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 991.705441] env[61867]: DEBUG oslo_concurrency.lockutils [req-41b7dd23-3ad2-46f5-a04f-adf0cc3c033b req-eff04d60-0f66-41af-81d6-ac62440ba2af service nova] Acquiring lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.744451] env[61867]: DEBUG nova.network.neutron [-] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.759165] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528276a7-e663-bdb4-396f-4abd1975e5e4, 'name': SearchDatastore_Task, 'duration_secs': 0.033776} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.761352] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.761480] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2/df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 991.763691] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9e6d3ecb-63a0-45b2-a783-1fa1c6c768ca {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.770736] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 991.770736] env[61867]: value = "task-1277035" [ 991.770736] env[61867]: _type = "Task" [ 991.770736] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.779992] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277035, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.820613] env[61867]: DEBUG oslo_concurrency.lockutils [req-5b819e60-4d70-441a-8c60-b8afda5a0a55 req-8008b0c6-432f-481a-a608-5c5cc76c2edb service nova] Releasing lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.833689] env[61867]: DEBUG oslo_concurrency.lockutils [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.834442] env[61867]: DEBUG oslo_concurrency.lockutils [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.834622] env[61867]: DEBUG oslo_concurrency.lockutils [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.835388] env[61867]: DEBUG oslo_concurrency.lockutils [req-41b7dd23-3ad2-46f5-a04f-adf0cc3c033b req-eff04d60-0f66-41af-81d6-ac62440ba2af service nova] Acquired lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.835605] env[61867]: DEBUG nova.network.neutron [req-41b7dd23-3ad2-46f5-a04f-adf0cc3c033b req-eff04d60-0f66-41af-81d6-ac62440ba2af service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Refreshing network info cache for port aa5b1005-fdcb-4208-b0d8-3b33573d05b6 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 991.837445] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a564f112-09b8-4dff-9227-8f32f208324c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.857566] env[61867]: DEBUG nova.virt.hardware [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 991.857842] env[61867]: DEBUG nova.virt.hardware [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 991.858015] env[61867]: DEBUG nova.virt.hardware [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 991.858211] env[61867]: DEBUG nova.virt.hardware [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 991.858363] env[61867]: DEBUG nova.virt.hardware [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 991.858515] env[61867]: DEBUG nova.virt.hardware [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 991.858729] env[61867]: DEBUG nova.virt.hardware [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 991.858916] env[61867]: DEBUG nova.virt.hardware [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 991.859096] env[61867]: DEBUG nova.virt.hardware [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 991.859273] env[61867]: DEBUG nova.virt.hardware [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 991.859471] env[61867]: DEBUG nova.virt.hardware [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 991.866154] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Reconfiguring VM to attach interface {{(pid=61867) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 991.866963] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-779bd05d-1278-4613-b077-83bd3e7082fb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.885954] env[61867]: DEBUG oslo_vmware.api [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 991.885954] env[61867]: value = "task-1277036" [ 991.885954] env[61867]: _type = "Task" [ 991.885954] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.895030] env[61867]: DEBUG oslo_vmware.api [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277036, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.925512] env[61867]: INFO nova.compute.manager [-] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Took 1.47 seconds to deallocate network for instance. [ 992.087616] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277034, 'name': Rename_Task, 'duration_secs': 0.147613} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.087972] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 992.088276] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5bde8f17-2441-44bd-917e-8244a4ff4777 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.097248] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 992.097248] env[61867]: value = "task-1277037" [ 992.097248] env[61867]: _type = "Task" [ 992.097248] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.110390] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277037, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.203455] env[61867]: DEBUG nova.network.neutron [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Updating instance_info_cache with network_info: [{"id": "e070a089-e0e7-4b11-9688-26d528b37e6a", "address": "fa:16:3e:7b:ac:a0", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape070a089-e0", "ovs_interfaceid": "e070a089-e0e7-4b11-9688-26d528b37e6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.248629] env[61867]: INFO nova.compute.manager [-] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Took 2.22 seconds to deallocate network for instance. [ 992.264633] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a9943bc8-1532-4b44-8a3e-068b4e0d1e46 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "976299f8-c6cb-4106-90f2-0a4c85625d8c" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.265141] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a9943bc8-1532-4b44-8a3e-068b4e0d1e46 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "976299f8-c6cb-4106-90f2-0a4c85625d8c" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.265444] env[61867]: DEBUG nova.compute.manager [None req-a9943bc8-1532-4b44-8a3e-068b4e0d1e46 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 992.266521] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d383f15-d697-43e7-bce9-ba457ee9c199 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.277687] env[61867]: DEBUG nova.compute.manager [None req-a9943bc8-1532-4b44-8a3e-068b4e0d1e46 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61867) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 992.278445] env[61867]: DEBUG nova.objects.instance [None req-a9943bc8-1532-4b44-8a3e-068b4e0d1e46 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lazy-loading 'flavor' on Instance uuid 976299f8-c6cb-4106-90f2-0a4c85625d8c {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.283626] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277035, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.396211] env[61867]: DEBUG oslo_vmware.api [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277036, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.431693] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.431855] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.432176] env[61867]: DEBUG nova.objects.instance [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Lazy-loading 'resources' on Instance uuid d19e7706-e6ac-47dd-8ec3-c30bd315559d {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.439163] env[61867]: DEBUG nova.compute.manager [req-5c0843bf-c4b5-406e-a20d-de5230418aec req-85255ae4-8709-46d4-9601-fd2864069006 service nova] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Received event network-vif-deleted-213271df-f382-4d6d-b1ec-2077126a4827 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.552522] env[61867]: DEBUG nova.network.neutron [req-41b7dd23-3ad2-46f5-a04f-adf0cc3c033b req-eff04d60-0f66-41af-81d6-ac62440ba2af service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Updated VIF entry in instance network info cache for port aa5b1005-fdcb-4208-b0d8-3b33573d05b6. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 992.553106] env[61867]: DEBUG nova.network.neutron [req-41b7dd23-3ad2-46f5-a04f-adf0cc3c033b req-eff04d60-0f66-41af-81d6-ac62440ba2af service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Updating instance_info_cache with network_info: [{"id": "c05702c4-5138-450d-82b2-790b86d8a59b", "address": "fa:16:3e:cf:d4:2a", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc05702c4-51", "ovs_interfaceid": "c05702c4-5138-450d-82b2-790b86d8a59b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "aa5b1005-fdcb-4208-b0d8-3b33573d05b6", "address": "fa:16:3e:80:74:6e", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa5b1005-fd", "ovs_interfaceid": "aa5b1005-fdcb-4208-b0d8-3b33573d05b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.611829] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277037, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.706415] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "refresh_cache-6e2136e7-c7e0-4a98-9899-f79c10f0e703" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.706705] env[61867]: DEBUG nova.objects.instance [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lazy-loading 'migration_context' on Instance uuid 6e2136e7-c7e0-4a98-9899-f79c10f0e703 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.757929] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.784176] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277035, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534503} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.786830] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2/df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 992.787138] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 992.787756] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9943bc8-1532-4b44-8a3e-068b4e0d1e46 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 992.788091] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-30516a6a-7b9d-4d5f-9b79-c1e5729fc0ee {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.790270] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-64b08efb-f4e0-4d87-a8ac-0c05cc178c51 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.796770] env[61867]: DEBUG oslo_vmware.api [None req-a9943bc8-1532-4b44-8a3e-068b4e0d1e46 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 992.796770] env[61867]: value = "task-1277039" [ 992.796770] env[61867]: _type = "Task" [ 992.796770] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.798153] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 992.798153] env[61867]: value = "task-1277038" [ 992.798153] env[61867]: _type = "Task" [ 992.798153] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.810473] env[61867]: DEBUG oslo_vmware.api [None req-a9943bc8-1532-4b44-8a3e-068b4e0d1e46 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1277039, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.814620] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277038, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.896503] env[61867]: DEBUG oslo_vmware.api [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277036, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.057138] env[61867]: DEBUG oslo_concurrency.lockutils [req-41b7dd23-3ad2-46f5-a04f-adf0cc3c033b req-eff04d60-0f66-41af-81d6-ac62440ba2af service nova] Releasing lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.057625] env[61867]: DEBUG nova.compute.manager [req-41b7dd23-3ad2-46f5-a04f-adf0cc3c033b req-eff04d60-0f66-41af-81d6-ac62440ba2af service nova] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Received event network-vif-deleted-a3a61d3e-16b8-476c-8ff4-fb229c392f72 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.057978] env[61867]: INFO nova.compute.manager [req-41b7dd23-3ad2-46f5-a04f-adf0cc3c033b req-eff04d60-0f66-41af-81d6-ac62440ba2af service nova] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Neutron deleted interface a3a61d3e-16b8-476c-8ff4-fb229c392f72; detaching it from the instance and deleting it from the info cache [ 993.058322] env[61867]: DEBUG nova.network.neutron [req-41b7dd23-3ad2-46f5-a04f-adf0cc3c033b req-eff04d60-0f66-41af-81d6-ac62440ba2af service nova] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.118319] env[61867]: DEBUG oslo_vmware.api [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277037, 'name': PowerOnVM_Task, 'duration_secs': 0.559509} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.122020] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 993.122020] env[61867]: INFO nova.compute.manager [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Took 8.01 seconds to spawn the instance on the hypervisor. [ 993.122020] env[61867]: DEBUG nova.compute.manager [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 993.122020] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a24be81f-6f03-4a73-928a-732a4d4b308c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.126548] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a4ea595-2392-43b7-8f94-2ae53326bbda {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.142395] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cecc637-b15d-4add-9098-6c72fe5a701f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.175046] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc00b1c-69a8-4cd7-a12e-721318a44a98 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.210206] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82e8780-3696-4b37-8741-8c34f0f301b9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.210206] env[61867]: DEBUG nova.compute.provider_tree [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 993.210206] env[61867]: DEBUG nova.objects.base [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Object Instance<6e2136e7-c7e0-4a98-9899-f79c10f0e703> lazy-loaded attributes: info_cache,migration_context {{(pid=61867) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 993.210893] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd3eaf6f-3bd4-4e7f-b0a0-54a6b32f4ddf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.233488] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa0afcb5-7d96-4bd5-a575-7c11f710ef0c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.239708] env[61867]: DEBUG oslo_vmware.api [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 993.239708] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528c7568-b8a5-8148-134b-d1e5d5e7a1f0" [ 993.239708] env[61867]: _type = "Task" [ 993.239708] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.249037] env[61867]: DEBUG oslo_vmware.api [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528c7568-b8a5-8148-134b-d1e5d5e7a1f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.310432] env[61867]: DEBUG oslo_vmware.api [None req-a9943bc8-1532-4b44-8a3e-068b4e0d1e46 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1277039, 'name': PowerOffVM_Task, 'duration_secs': 0.372025} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.310743] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9943bc8-1532-4b44-8a3e-068b4e0d1e46 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 993.310922] env[61867]: DEBUG nova.compute.manager [None req-a9943bc8-1532-4b44-8a3e-068b4e0d1e46 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 993.311955] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e967fd8-66b0-4b1d-820d-4c2ff1bd3dea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.316901] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277038, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069254} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.317429] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 993.318133] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0277f254-c318-4acc-9119-8bdbfab766e7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.341774] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2/df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 993.342019] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa3f0024-5173-45c2-a506-882c3d8e6033 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.362858] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 993.362858] env[61867]: value = "task-1277040" [ 993.362858] env[61867]: _type = "Task" [ 993.362858] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.370987] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277040, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.396941] env[61867]: DEBUG oslo_vmware.api [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277036, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.562069] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b8a95a19-a784-4a0c-909c-b909a704b79b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.570951] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b87a3f-560c-4ca0-8174-5950c4502280 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.597273] env[61867]: DEBUG nova.compute.manager [req-41b7dd23-3ad2-46f5-a04f-adf0cc3c033b req-eff04d60-0f66-41af-81d6-ac62440ba2af service nova] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Detach interface failed, port_id=a3a61d3e-16b8-476c-8ff4-fb229c392f72, reason: Instance 2aa08603-d87f-4734-bdfe-fdd610d54e1f could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 993.649388] env[61867]: INFO nova.compute.manager [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Took 21.24 seconds to build instance. [ 993.722588] env[61867]: ERROR nova.scheduler.client.report [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] [req-ff73059a-0344-49b0-99d2-a3e29704cf6c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 25720271-a549-4916-abe3-e5ed9b765889. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ff73059a-0344-49b0-99d2-a3e29704cf6c"}]} [ 993.739085] env[61867]: DEBUG nova.scheduler.client.report [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Refreshing inventories for resource provider 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 993.749961] env[61867]: DEBUG oslo_vmware.api [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528c7568-b8a5-8148-134b-d1e5d5e7a1f0, 'name': SearchDatastore_Task, 'duration_secs': 0.007494} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.750271] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.753467] env[61867]: DEBUG nova.scheduler.client.report [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Updating ProviderTree inventory for provider 25720271-a549-4916-abe3-e5ed9b765889 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 993.753694] env[61867]: DEBUG nova.compute.provider_tree [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 993.764558] env[61867]: DEBUG nova.scheduler.client.report [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Refreshing aggregate associations for resource provider 25720271-a549-4916-abe3-e5ed9b765889, aggregates: None {{(pid=61867) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 993.781798] env[61867]: DEBUG nova.scheduler.client.report [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Refreshing trait associations for resource provider 25720271-a549-4916-abe3-e5ed9b765889, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61867) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 993.827205] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a9943bc8-1532-4b44-8a3e-068b4e0d1e46 tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "976299f8-c6cb-4106-90f2-0a4c85625d8c" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.562s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.875482] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277040, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.898593] env[61867]: DEBUG oslo_vmware.api [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277036, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.922607] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1562ac51-7ea6-49d9-a1e9-672bb14dedce {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.929928] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb0cf08-f681-43cb-bc45-71e036fb6594 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.961648] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac0e52b8-d502-4745-a27c-c338313a39b2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.969268] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db038630-9398-4bbb-b36c-784578cec626 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.983311] env[61867]: DEBUG nova.compute.provider_tree [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 994.151640] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a5ad49bb-4463-42b8-8f4d-031fd74c4b99 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "4006cabd-cb6e-45b6-b87e-9b4b1550f4fe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.757s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.375356] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277040, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.399246] env[61867]: DEBUG oslo_vmware.api [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277036, 'name': ReconfigVM_Task, 'duration_secs': 2.2303} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.399571] env[61867]: DEBUG oslo_concurrency.lockutils [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.399790] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Reconfigured VM to attach interface {{(pid=61867) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 994.516954] env[61867]: DEBUG nova.scheduler.client.report [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Updated inventory for provider 25720271-a549-4916-abe3-e5ed9b765889 with generation 124 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 994.517257] env[61867]: DEBUG nova.compute.provider_tree [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Updating resource provider 25720271-a549-4916-abe3-e5ed9b765889 generation from 124 to 125 during operation: update_inventory {{(pid=61867) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 994.517444] env[61867]: DEBUG nova.compute.provider_tree [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Updating inventory in ProviderTree for provider 25720271-a549-4916-abe3-e5ed9b765889 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 994.567984] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e600870-9d53-4f44-88c8-45f114de1a6b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.573889] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-5302adc0-978c-48e4-b626-069ad8f43859 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Suspending the VM {{(pid=61867) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 994.574148] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-81817587-4e72-45b1-af6f-c77b745d6874 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.579707] env[61867]: DEBUG oslo_vmware.api [None req-5302adc0-978c-48e4-b626-069ad8f43859 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 994.579707] env[61867]: value = "task-1277041" [ 994.579707] env[61867]: _type = "Task" [ 994.579707] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.586904] env[61867]: DEBUG oslo_vmware.api [None req-5302adc0-978c-48e4-b626-069ad8f43859 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277041, 'name': SuspendVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.842242] env[61867]: DEBUG oslo_concurrency.lockutils [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "8a83f4f2-58eb-473e-9b1e-32ce633554f9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.842674] env[61867]: DEBUG oslo_concurrency.lockutils [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "8a83f4f2-58eb-473e-9b1e-32ce633554f9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.842949] env[61867]: DEBUG oslo_concurrency.lockutils [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "8a83f4f2-58eb-473e-9b1e-32ce633554f9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.843282] env[61867]: DEBUG oslo_concurrency.lockutils [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "8a83f4f2-58eb-473e-9b1e-32ce633554f9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.843602] env[61867]: DEBUG oslo_concurrency.lockutils [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "8a83f4f2-58eb-473e-9b1e-32ce633554f9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.846524] env[61867]: INFO nova.compute.manager [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Terminating instance [ 994.848961] env[61867]: DEBUG nova.compute.manager [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 994.849207] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 994.850294] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c44c68-378d-43d7-a16d-6a3c43c225a0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.860910] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 994.861331] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74a756b9-73e1-4dcf-8643-1ef6eb7f40c1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.870737] env[61867]: DEBUG oslo_vmware.api [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 994.870737] env[61867]: value = "task-1277042" [ 994.870737] env[61867]: _type = "Task" [ 994.870737] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.881644] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277040, 'name': ReconfigVM_Task, 'duration_secs': 1.509509} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.882724] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Reconfigured VM instance instance-00000064 to attach disk [datastore1] df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2/df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 994.888941] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-61c2fbeb-4bc9-4ca7-bcb5-aba62dfc85e6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.891178] env[61867]: DEBUG oslo_vmware.api [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1277042, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.897270] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 994.897270] env[61867]: value = "task-1277043" [ 994.897270] env[61867]: _type = "Task" [ 994.897270] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.904550] env[61867]: DEBUG oslo_concurrency.lockutils [None req-009bf89d-5cc5-438c-b7cc-abe2d8ec0c2f tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "interface-4d7bfb31-d565-49e4-8c51-0122acebba2d-aa5b1005-fdcb-4208-b0d8-3b33573d05b6" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.446s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.915040] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277043, 'name': Rename_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.022773] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.591s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.025907] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.268s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.026484] env[61867]: DEBUG nova.objects.instance [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lazy-loading 'resources' on Instance uuid 2aa08603-d87f-4734-bdfe-fdd610d54e1f {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.048773] env[61867]: INFO nova.scheduler.client.report [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Deleted allocations for instance d19e7706-e6ac-47dd-8ec3-c30bd315559d [ 995.091984] env[61867]: DEBUG oslo_vmware.api [None req-5302adc0-978c-48e4-b626-069ad8f43859 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277041, 'name': SuspendVM_Task} progress is 70%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.381535] env[61867]: DEBUG oslo_vmware.api [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1277042, 'name': PowerOffVM_Task, 'duration_secs': 0.241974} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.381846] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 995.382116] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 995.382347] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0ac8d240-70d5-49e3-9ad1-04920efaaf66 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.406848] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277043, 'name': Rename_Task, 'duration_secs': 0.18695} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.407151] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 995.407449] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5e8273d-4ae0-4f3c-86e4-e8a109fa03d3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.414902] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 995.414902] env[61867]: value = "task-1277045" [ 995.414902] env[61867]: _type = "Task" [ 995.414902] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.422764] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277045, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.528599] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 995.528890] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 995.529039] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Deleting the datastore file [datastore2] 8a83f4f2-58eb-473e-9b1e-32ce633554f9 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 995.529346] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c8a21f6a-5d4a-46df-bf60-23825c4fa4dc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.538853] env[61867]: DEBUG oslo_vmware.api [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 995.538853] env[61867]: value = "task-1277046" [ 995.538853] env[61867]: _type = "Task" [ 995.538853] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.546736] env[61867]: DEBUG oslo_vmware.api [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1277046, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.556344] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caf077e1-e247-41d5-9258-376ae234fb2a tempest-AttachInterfacesV270Test-119640540 tempest-AttachInterfacesV270Test-119640540-project-member] Lock "d19e7706-e6ac-47dd-8ec3-c30bd315559d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.746s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.594028] env[61867]: DEBUG oslo_vmware.api [None req-5302adc0-978c-48e4-b626-069ad8f43859 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277041, 'name': SuspendVM_Task, 'duration_secs': 0.626076} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.594150] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-5302adc0-978c-48e4-b626-069ad8f43859 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Suspended the VM {{(pid=61867) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 995.594333] env[61867]: DEBUG nova.compute.manager [None req-5302adc0-978c-48e4-b626-069ad8f43859 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 995.595143] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9282f14e-d5ae-448c-99ac-81b5d70b13ab {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.679093] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7073f16-7929-45f5-8442-18bc94fe84d3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.687654] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17359c17-5dcc-4c7b-9753-ce168b2cceab {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.720576] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4fd89d-47d8-4092-8cdd-1aeeea90fcf6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.730806] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc5689d-d586-445c-9dda-f742d54c0498 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.744572] env[61867]: DEBUG nova.compute.provider_tree [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.924989] env[61867]: DEBUG oslo_vmware.api [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277045, 'name': PowerOnVM_Task, 'duration_secs': 0.481593} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.924989] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 995.924989] env[61867]: INFO nova.compute.manager [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Took 8.52 seconds to spawn the instance on the hypervisor. [ 995.924989] env[61867]: DEBUG nova.compute.manager [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 995.925718] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6454352a-fe34-4694-bfb8-cf6a5451961c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.055212] env[61867]: DEBUG oslo_vmware.api [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1277046, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191934} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.055212] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 996.055212] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 996.055212] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 996.055212] env[61867]: INFO nova.compute.manager [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Took 1.20 seconds to destroy the instance on the hypervisor. [ 996.055212] env[61867]: DEBUG oslo.service.loopingcall [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.055212] env[61867]: DEBUG nova.compute.manager [-] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 996.055212] env[61867]: DEBUG nova.network.neutron [-] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 996.147609] env[61867]: DEBUG oslo_concurrency.lockutils [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "976299f8-c6cb-4106-90f2-0a4c85625d8c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.147609] env[61867]: DEBUG oslo_concurrency.lockutils [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "976299f8-c6cb-4106-90f2-0a4c85625d8c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.147609] env[61867]: DEBUG oslo_concurrency.lockutils [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "976299f8-c6cb-4106-90f2-0a4c85625d8c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.147609] env[61867]: DEBUG oslo_concurrency.lockutils [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "976299f8-c6cb-4106-90f2-0a4c85625d8c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.147609] env[61867]: DEBUG oslo_concurrency.lockutils [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "976299f8-c6cb-4106-90f2-0a4c85625d8c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.148331] env[61867]: INFO nova.compute.manager [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Terminating instance [ 996.150280] env[61867]: DEBUG nova.compute.manager [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 996.150589] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 996.151634] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db03cd48-c2c0-4e7f-bbc8-da43f509ce61 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.160769] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 996.161123] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f79f4bd-ecc5-4dcf-add2-24566816c515 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.175158] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "interface-4d7bfb31-d565-49e4-8c51-0122acebba2d-aa5b1005-fdcb-4208-b0d8-3b33573d05b6" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.175562] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "interface-4d7bfb31-d565-49e4-8c51-0122acebba2d-aa5b1005-fdcb-4208-b0d8-3b33573d05b6" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.223612] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 996.224106] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 996.224400] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleting the datastore file [datastore2] 976299f8-c6cb-4106-90f2-0a4c85625d8c {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 996.224757] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11ebad66-435c-473d-b36b-329afb03f82f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.233019] env[61867]: DEBUG oslo_vmware.api [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for the task: (returnval){ [ 996.233019] env[61867]: value = "task-1277048" [ 996.233019] env[61867]: _type = "Task" [ 996.233019] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.239745] env[61867]: DEBUG oslo_vmware.api [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1277048, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.246807] env[61867]: DEBUG nova.scheduler.client.report [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.442180] env[61867]: INFO nova.compute.manager [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Took 17.43 seconds to build instance. [ 996.504584] env[61867]: DEBUG nova.compute.manager [req-40b2e567-ac7f-4d51-a9c8-1ca7b9a6c655 req-30348612-c633-4969-8747-da497efcc49e service nova] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Received event network-vif-deleted-cdf39635-188f-4031-ae51-c7d98a8e8365 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 996.504813] env[61867]: INFO nova.compute.manager [req-40b2e567-ac7f-4d51-a9c8-1ca7b9a6c655 req-30348612-c633-4969-8747-da497efcc49e service nova] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Neutron deleted interface cdf39635-188f-4031-ae51-c7d98a8e8365; detaching it from the instance and deleting it from the info cache [ 996.505591] env[61867]: DEBUG nova.network.neutron [req-40b2e567-ac7f-4d51-a9c8-1ca7b9a6c655 req-30348612-c633-4969-8747-da497efcc49e service nova] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.681912] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.682598] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.685324] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef41aa0-1906-4de0-8f61-d645a43d1fe8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.709311] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d8e0ba9-cf24-4b37-80dc-456629bd076a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.741023] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Reconfiguring VM to detach interface {{(pid=61867) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 996.746095] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fde0981a-9961-42cb-8626-e2c789f13b20 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.760129] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.734s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.762927] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 3.013s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.773526] env[61867]: DEBUG oslo_vmware.api [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Task: {'id': task-1277048, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.446457} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.776748] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 996.776748] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 996.776849] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 996.777056] env[61867]: INFO nova.compute.manager [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Took 0.63 seconds to destroy the instance on the hypervisor. [ 996.777323] env[61867]: DEBUG oslo.service.loopingcall [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.778387] env[61867]: DEBUG oslo_vmware.api [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 996.778387] env[61867]: value = "task-1277049" [ 996.778387] env[61867]: _type = "Task" [ 996.778387] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.778860] env[61867]: DEBUG nova.compute.manager [-] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 996.778860] env[61867]: DEBUG nova.network.neutron [-] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 996.792059] env[61867]: DEBUG oslo_vmware.api [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.792059] env[61867]: INFO nova.scheduler.client.report [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Deleted allocations for instance 2aa08603-d87f-4734-bdfe-fdd610d54e1f [ 996.945908] env[61867]: DEBUG oslo_concurrency.lockutils [None req-1a73881b-1cd9-47e5-8bc1-f40105341e8e tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.944s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.977802] env[61867]: DEBUG nova.network.neutron [-] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.011023] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-23f11fd7-76a6-45ca-9311-4338996740aa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.025178] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9583455c-9728-4065-aae1-f54b1844c9df {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.050589] env[61867]: DEBUG nova.compute.manager [req-40b2e567-ac7f-4d51-a9c8-1ca7b9a6c655 req-30348612-c633-4969-8747-da497efcc49e service nova] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Detach interface failed, port_id=cdf39635-188f-4031-ae51-c7d98a8e8365, reason: Instance 8a83f4f2-58eb-473e-9b1e-32ce633554f9 could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 997.183084] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "4006cabd-cb6e-45b6-b87e-9b4b1550f4fe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.183657] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "4006cabd-cb6e-45b6-b87e-9b4b1550f4fe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.183999] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "4006cabd-cb6e-45b6-b87e-9b4b1550f4fe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.184266] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "4006cabd-cb6e-45b6-b87e-9b4b1550f4fe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.184486] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "4006cabd-cb6e-45b6-b87e-9b4b1550f4fe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.186828] env[61867]: INFO nova.compute.manager [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Terminating instance [ 997.189014] env[61867]: DEBUG nova.compute.manager [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 997.189271] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 997.190121] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e61a3dd-62c1-48e6-9dc9-e81cf8581159 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.200796] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 997.202202] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-47145e4d-fd92-4609-89ce-e266ff144076 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.277311] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 997.278912] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 997.279289] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleting the datastore file [datastore1] 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 997.279561] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-612e2822-35c9-450a-9134-e4d5194a884c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.293981] env[61867]: DEBUG oslo_vmware.api [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.296968] env[61867]: DEBUG oslo_vmware.api [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 997.296968] env[61867]: value = "task-1277051" [ 997.296968] env[61867]: _type = "Task" [ 997.296968] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.306102] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8d69e5f1-864a-4673-926d-822d37d201f9 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "2aa08603-d87f-4734-bdfe-fdd610d54e1f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.421s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.314824] env[61867]: DEBUG oslo_vmware.api [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277051, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.331009] env[61867]: DEBUG nova.compute.manager [req-0a4f3e4b-7b9a-4ddb-bc2f-985c64a0f65c req-fb2357aa-672c-45e0-a7bc-68b4eece287b service nova] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Received event network-changed-f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.332727] env[61867]: DEBUG nova.compute.manager [req-0a4f3e4b-7b9a-4ddb-bc2f-985c64a0f65c req-fb2357aa-672c-45e0-a7bc-68b4eece287b service nova] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Refreshing instance network info cache due to event network-changed-f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 997.332727] env[61867]: DEBUG oslo_concurrency.lockutils [req-0a4f3e4b-7b9a-4ddb-bc2f-985c64a0f65c req-fb2357aa-672c-45e0-a7bc-68b4eece287b service nova] Acquiring lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.332727] env[61867]: DEBUG oslo_concurrency.lockutils [req-0a4f3e4b-7b9a-4ddb-bc2f-985c64a0f65c req-fb2357aa-672c-45e0-a7bc-68b4eece287b service nova] Acquired lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.332727] env[61867]: DEBUG nova.network.neutron [req-0a4f3e4b-7b9a-4ddb-bc2f-985c64a0f65c req-fb2357aa-672c-45e0-a7bc-68b4eece287b service nova] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Refreshing network info cache for port f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 997.434023] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1679ee0-757f-476a-aea3-ae9f2a27792c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.444728] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d9164a1-656b-4113-af92-8ba2ae155743 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.480154] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2887f486-0967-4593-85d0-d376562ace27 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.484160] env[61867]: INFO nova.compute.manager [-] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Took 1.43 seconds to deallocate network for instance. [ 997.498928] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76aa1e57-1925-4362-a6ce-f8899a2e27cb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.514410] env[61867]: DEBUG nova.compute.provider_tree [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.717726] env[61867]: DEBUG nova.network.neutron [-] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.794658] env[61867]: DEBUG oslo_vmware.api [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.810690] env[61867]: DEBUG oslo_vmware.api [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277051, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15198} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.811038] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 997.811198] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 997.811999] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 997.811999] env[61867]: INFO nova.compute.manager [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Took 0.62 seconds to destroy the instance on the hypervisor. [ 997.811999] env[61867]: DEBUG oslo.service.loopingcall [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 997.811999] env[61867]: DEBUG nova.compute.manager [-] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 997.812208] env[61867]: DEBUG nova.network.neutron [-] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 997.992777] env[61867]: DEBUG oslo_concurrency.lockutils [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.020736] env[61867]: DEBUG nova.scheduler.client.report [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.193480] env[61867]: DEBUG nova.network.neutron [req-0a4f3e4b-7b9a-4ddb-bc2f-985c64a0f65c req-fb2357aa-672c-45e0-a7bc-68b4eece287b service nova] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updated VIF entry in instance network info cache for port f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 998.193862] env[61867]: DEBUG nova.network.neutron [req-0a4f3e4b-7b9a-4ddb-bc2f-985c64a0f65c req-fb2357aa-672c-45e0-a7bc-68b4eece287b service nova] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating instance_info_cache with network_info: [{"id": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "address": "fa:16:3e:b9:62:3b", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1e84fbf-9f", "ovs_interfaceid": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.219935] env[61867]: INFO nova.compute.manager [-] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Took 1.44 seconds to deallocate network for instance. [ 998.293073] env[61867]: DEBUG oslo_vmware.api [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.540806] env[61867]: DEBUG nova.compute.manager [req-79e775d9-0690-4dad-8c1a-293f809b1dcb req-6f177c27-be64-4070-8a2c-d911824f94d1 service nova] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Received event network-vif-deleted-92925b9b-00ea-4282-9781-7e60212ae093 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 998.541396] env[61867]: DEBUG nova.compute.manager [req-79e775d9-0690-4dad-8c1a-293f809b1dcb req-6f177c27-be64-4070-8a2c-d911824f94d1 service nova] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Received event network-vif-deleted-c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 998.541632] env[61867]: INFO nova.compute.manager [req-79e775d9-0690-4dad-8c1a-293f809b1dcb req-6f177c27-be64-4070-8a2c-d911824f94d1 service nova] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Neutron deleted interface c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78; detaching it from the instance and deleting it from the info cache [ 998.541820] env[61867]: DEBUG nova.network.neutron [req-79e775d9-0690-4dad-8c1a-293f809b1dcb req-6f177c27-be64-4070-8a2c-d911824f94d1 service nova] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.699167] env[61867]: DEBUG oslo_concurrency.lockutils [req-0a4f3e4b-7b9a-4ddb-bc2f-985c64a0f65c req-fb2357aa-672c-45e0-a7bc-68b4eece287b service nova] Releasing lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.713851] env[61867]: DEBUG nova.network.neutron [-] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.725874] env[61867]: DEBUG oslo_concurrency.lockutils [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.794085] env[61867]: DEBUG oslo_vmware.api [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.032343] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.269s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.035225] env[61867]: DEBUG oslo_concurrency.lockutils [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.043s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.035454] env[61867]: DEBUG nova.objects.instance [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lazy-loading 'resources' on Instance uuid 8a83f4f2-58eb-473e-9b1e-32ce633554f9 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.044102] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3e978d7-d60f-4c3c-be30-6ef479d2a457 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.054710] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1890b91e-ea03-4e2a-8029-2d9fe247efd6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.081946] env[61867]: DEBUG nova.compute.manager [req-79e775d9-0690-4dad-8c1a-293f809b1dcb req-6f177c27-be64-4070-8a2c-d911824f94d1 service nova] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Detach interface failed, port_id=c5ee9c92-b2da-4b06-ac12-ddc52bf6ac78, reason: Instance 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 999.216333] env[61867]: INFO nova.compute.manager [-] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Took 1.40 seconds to deallocate network for instance. [ 999.293217] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "748e8a71-28b6-4c91-9cbb-856359348c96" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.293430] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "748e8a71-28b6-4c91-9cbb-856359348c96" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.297982] env[61867]: DEBUG oslo_vmware.api [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.591570] env[61867]: INFO nova.scheduler.client.report [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleted allocation for migration 44205de9-d7fc-454a-8217-47bec978b3d8 [ 999.665952] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f22a7e-282d-4996-9849-42c0cce0e146 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.674719] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df5f79ed-6392-4d31-b7a6-a795532fa2db {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.704450] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e976a3a-8ae0-4b20-b7da-f70b7310f172 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.712872] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b747acf-470f-4578-8567-f21d3de74789 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.727880] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.728388] env[61867]: DEBUG nova.compute.provider_tree [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 999.796787] env[61867]: DEBUG oslo_vmware.api [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.799286] env[61867]: DEBUG nova.compute.manager [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 999.956525] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 999.956731] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1000.100796] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bad8948e-4059-46e0-a8fb-58cec859430c tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.879s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.156414] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.156697] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.156920] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.157131] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.157309] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.159422] env[61867]: INFO nova.compute.manager [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Terminating instance [ 1000.161563] env[61867]: DEBUG nova.compute.manager [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1000.161768] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1000.162617] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff4ccb87-9ef3-45d7-b30c-188acc2d73db {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.170345] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1000.170591] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54a2e0f0-10f5-46fa-9415-428b5d6f8626 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.178619] env[61867]: DEBUG oslo_vmware.api [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 1000.178619] env[61867]: value = "task-1277052" [ 1000.178619] env[61867]: _type = "Task" [ 1000.178619] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.187554] env[61867]: DEBUG oslo_vmware.api [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277052, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.231244] env[61867]: DEBUG nova.scheduler.client.report [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1000.295008] env[61867]: DEBUG oslo_vmware.api [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.321445] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.610925] env[61867]: DEBUG oslo_concurrency.lockutils [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "29d961df-5db5-46db-ace9-651913e0250a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.611210] env[61867]: DEBUG oslo_concurrency.lockutils [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "29d961df-5db5-46db-ace9-651913e0250a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.688587] env[61867]: DEBUG oslo_vmware.api [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277052, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.736633] env[61867]: DEBUG oslo_concurrency.lockutils [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.701s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.738782] env[61867]: DEBUG oslo_concurrency.lockutils [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.013s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.739034] env[61867]: DEBUG nova.objects.instance [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lazy-loading 'resources' on Instance uuid 976299f8-c6cb-4106-90f2-0a4c85625d8c {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.753972] env[61867]: INFO nova.scheduler.client.report [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Deleted allocations for instance 8a83f4f2-58eb-473e-9b1e-32ce633554f9 [ 1000.795318] env[61867]: DEBUG oslo_vmware.api [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.114176] env[61867]: DEBUG nova.compute.manager [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1001.189080] env[61867]: DEBUG oslo_vmware.api [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277052, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.263700] env[61867]: DEBUG oslo_concurrency.lockutils [None req-873d4f13-1394-45a8-b848-fafa6f4846a5 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "8a83f4f2-58eb-473e-9b1e-32ce633554f9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.421s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.297733] env[61867]: DEBUG oslo_vmware.api [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.366759] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-009139e6-8a39-4cf8-8662-9b68e8cb03df {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.374578] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a3dd2a-1229-4d0a-9b4a-a4f5de4bf4f3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.402911] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e096865c-3c16-486d-9f56-6ab85c04af5e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.410098] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9627e9e6-c233-41c5-bffd-0e742f179d63 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.422827] env[61867]: DEBUG nova.compute.provider_tree [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.631992] env[61867]: DEBUG oslo_concurrency.lockutils [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.667770] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "6e41989e-b8fa-4009-af1e-1ce859b329a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.668063] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "6e41989e-b8fa-4009-af1e-1ce859b329a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.668289] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "6e41989e-b8fa-4009-af1e-1ce859b329a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.668483] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "6e41989e-b8fa-4009-af1e-1ce859b329a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.668666] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "6e41989e-b8fa-4009-af1e-1ce859b329a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.671400] env[61867]: INFO nova.compute.manager [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Terminating instance [ 1001.673287] env[61867]: DEBUG nova.compute.manager [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1001.673492] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1001.674336] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b84482-d783-400d-a211-911fe41ede7e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.682024] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1001.685050] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cfe6986e-1edb-46f9-b592-5a254a5098d6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.691319] env[61867]: DEBUG oslo_vmware.api [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277052, 'name': PowerOffVM_Task, 'duration_secs': 1.245495} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.692366] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1001.692600] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1001.692867] env[61867]: DEBUG oslo_vmware.api [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 1001.692867] env[61867]: value = "task-1277053" [ 1001.692867] env[61867]: _type = "Task" [ 1001.692867] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.693071] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bd787945-0d18-46d3-881f-04662922434e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.702993] env[61867]: DEBUG oslo_vmware.api [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1277053, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.767860] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1001.768097] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1001.768287] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleting the datastore file [datastore2] 6e2136e7-c7e0-4a98-9899-f79c10f0e703 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.768565] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-325d5ee6-ebd0-4386-9201-ee1d8a144b4a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.774656] env[61867]: DEBUG oslo_vmware.api [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 1001.774656] env[61867]: value = "task-1277055" [ 1001.774656] env[61867]: _type = "Task" [ 1001.774656] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.781928] env[61867]: DEBUG oslo_vmware.api [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277055, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.793094] env[61867]: DEBUG oslo_vmware.api [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.926683] env[61867]: DEBUG nova.scheduler.client.report [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1001.957942] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1001.958123] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Starting heal instance info cache {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1002.204507] env[61867]: DEBUG oslo_vmware.api [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1277053, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.284797] env[61867]: DEBUG oslo_vmware.api [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277055, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.189039} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.285102] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.285302] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1002.285490] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1002.285672] env[61867]: INFO nova.compute.manager [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Took 2.12 seconds to destroy the instance on the hypervisor. [ 1002.285925] env[61867]: DEBUG oslo.service.loopingcall [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.286135] env[61867]: DEBUG nova.compute.manager [-] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1002.286231] env[61867]: DEBUG nova.network.neutron [-] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1002.296187] env[61867]: DEBUG oslo_vmware.api [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277049, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.432709] env[61867]: DEBUG oslo_concurrency.lockutils [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.694s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.436307] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.708s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.436307] env[61867]: DEBUG nova.objects.instance [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lazy-loading 'resources' on Instance uuid 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1002.460449] env[61867]: INFO nova.scheduler.client.report [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Deleted allocations for instance 976299f8-c6cb-4106-90f2-0a4c85625d8c [ 1002.463388] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Skipping network cache update for instance because it is being deleted. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9977}} [ 1002.547382] env[61867]: DEBUG nova.compute.manager [req-8c0036a6-e12f-4f5f-abfb-8f153a024c1d req-de48a543-e9e7-4755-8790-59fd127373be service nova] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Received event network-vif-deleted-e070a089-e0e7-4b11-9688-26d528b37e6a {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.547615] env[61867]: INFO nova.compute.manager [req-8c0036a6-e12f-4f5f-abfb-8f153a024c1d req-de48a543-e9e7-4755-8790-59fd127373be service nova] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Neutron deleted interface e070a089-e0e7-4b11-9688-26d528b37e6a; detaching it from the instance and deleting it from the info cache [ 1002.547805] env[61867]: DEBUG nova.network.neutron [req-8c0036a6-e12f-4f5f-abfb-8f153a024c1d req-de48a543-e9e7-4755-8790-59fd127373be service nova] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.704840] env[61867]: DEBUG oslo_vmware.api [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1277053, 'name': PowerOffVM_Task, 'duration_secs': 0.81917} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.705198] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1002.705368] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1002.705615] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1013f1b2-83e2-464a-828a-2b93b61a2bf6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.772009] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1002.772218] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1002.772415] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Deleting the datastore file [datastore1] 6e41989e-b8fa-4009-af1e-1ce859b329a1 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1002.772747] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1aa5e7fa-793c-43ca-bf73-c1c1d970f276 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.779472] env[61867]: DEBUG oslo_vmware.api [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for the task: (returnval){ [ 1002.779472] env[61867]: value = "task-1277057" [ 1002.779472] env[61867]: _type = "Task" [ 1002.779472] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.786805] env[61867]: DEBUG oslo_vmware.api [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1277057, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.795191] env[61867]: DEBUG oslo_vmware.api [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277049, 'name': ReconfigVM_Task, 'duration_secs': 5.754343} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.795415] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.795626] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Reconfigured VM to detach interface {{(pid=61867) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1002.971095] env[61867]: DEBUG oslo_concurrency.lockutils [None req-812815c8-7895-4c9b-ba6f-633839613b6b tempest-ServersTestJSON-1355820659 tempest-ServersTestJSON-1355820659-project-member] Lock "976299f8-c6cb-4106-90f2-0a4c85625d8c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.826s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.026751] env[61867]: DEBUG nova.network.neutron [-] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.051756] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-21abbf0d-c4bd-4dba-9444-036abb3a0ef9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.061086] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c656e48a-b997-42ae-90d5-5edd8caa0604 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.073221] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b215f0a-22f9-4e91-9efc-3bcac33f0b6c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.091589] env[61867]: DEBUG nova.compute.manager [req-8c0036a6-e12f-4f5f-abfb-8f153a024c1d req-de48a543-e9e7-4755-8790-59fd127373be service nova] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Detach interface failed, port_id=e070a089-e0e7-4b11-9688-26d528b37e6a, reason: Instance 6e2136e7-c7e0-4a98-9899-f79c10f0e703 could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1003.096165] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96850a50-b3a0-4a7d-9051-617325c16673 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.136783] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3e370ed-41a0-4dd1-a7e7-c0ea95ec31d6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.146501] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fea858-ef2e-458e-940b-99cb7466f699 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.160027] env[61867]: DEBUG nova.compute.provider_tree [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1003.291756] env[61867]: DEBUG oslo_vmware.api [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Task: {'id': task-1277057, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134294} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.293264] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1003.293515] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1003.293717] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1003.293901] env[61867]: INFO nova.compute.manager [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Took 1.62 seconds to destroy the instance on the hypervisor. [ 1003.294166] env[61867]: DEBUG oslo.service.loopingcall [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.294367] env[61867]: DEBUG nova.compute.manager [-] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1003.294462] env[61867]: DEBUG nova.network.neutron [-] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1003.529905] env[61867]: INFO nova.compute.manager [-] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Took 1.24 seconds to deallocate network for instance. [ 1003.665651] env[61867]: DEBUG nova.scheduler.client.report [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1004.041431] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.172256] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.737s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.175442] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.854s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.176809] env[61867]: INFO nova.compute.claims [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1004.207182] env[61867]: INFO nova.scheduler.client.report [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleted allocations for instance 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe [ 1004.237493] env[61867]: DEBUG nova.network.neutron [-] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.271070] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.271441] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.271732] env[61867]: DEBUG nova.network.neutron [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1004.719280] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4ec8002d-8ed4-49ce-a9a4-50c50410dc19 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "4006cabd-cb6e-45b6-b87e-9b4b1550f4fe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.536s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.739934] env[61867]: INFO nova.compute.manager [-] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Took 1.45 seconds to deallocate network for instance. [ 1004.812226] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "interface-0ae83268-ef50-42e6-954d-5e67c4afc20e-aa5b1005-fdcb-4208-b0d8-3b33573d05b6" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.812226] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "interface-0ae83268-ef50-42e6-954d-5e67c4afc20e-aa5b1005-fdcb-4208-b0d8-3b33573d05b6" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.812226] env[61867]: DEBUG nova.objects.instance [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lazy-loading 'flavor' on Instance uuid 0ae83268-ef50-42e6-954d-5e67c4afc20e {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1004.917154] env[61867]: DEBUG nova.compute.manager [req-f5fe344d-bfda-4b67-a8e4-95a52c767ad0 req-867bc348-2459-4cbf-9f01-1bb6c0dede04 service nova] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Received event network-vif-deleted-d8069fac-960f-49cf-a416-036e26060a27 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.917400] env[61867]: DEBUG nova.compute.manager [req-f5fe344d-bfda-4b67-a8e4-95a52c767ad0 req-867bc348-2459-4cbf-9f01-1bb6c0dede04 service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Received event network-changed-c05702c4-5138-450d-82b2-790b86d8a59b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.917575] env[61867]: DEBUG nova.compute.manager [req-f5fe344d-bfda-4b67-a8e4-95a52c767ad0 req-867bc348-2459-4cbf-9f01-1bb6c0dede04 service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Refreshing instance network info cache due to event network-changed-c05702c4-5138-450d-82b2-790b86d8a59b. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1004.917855] env[61867]: DEBUG oslo_concurrency.lockutils [req-f5fe344d-bfda-4b67-a8e4-95a52c767ad0 req-867bc348-2459-4cbf-9f01-1bb6c0dede04 service nova] Acquiring lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.022811] env[61867]: INFO nova.network.neutron [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Port aa5b1005-fdcb-4208-b0d8-3b33573d05b6 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1005.023200] env[61867]: DEBUG nova.network.neutron [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Updating instance_info_cache with network_info: [{"id": "c05702c4-5138-450d-82b2-790b86d8a59b", "address": "fa:16:3e:cf:d4:2a", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc05702c4-51", "ovs_interfaceid": "c05702c4-5138-450d-82b2-790b86d8a59b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.247082] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.334054] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-747eca09-66b2-4d85-b863-6cddc085e546 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.341577] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab76b73-42f3-48cf-bac7-437c1bafd383 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.377231] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e7c70b9-7554-4b14-821b-ab41128f59e9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.385107] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca22c8ab-88ab-4787-a280-a27c724d8778 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.400303] env[61867]: DEBUG nova.compute.provider_tree [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1005.428713] env[61867]: DEBUG nova.objects.instance [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lazy-loading 'pci_requests' on Instance uuid 0ae83268-ef50-42e6-954d-5e67c4afc20e {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.482967] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Didn't find any instances for network info cache update. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1005.483084] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1005.483343] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1005.483676] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1005.483883] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61867) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1005.484133] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1005.521542] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "986ada58-eb26-4289-bc0e-338c6a399645" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.521788] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "986ada58-eb26-4289-bc0e-338c6a399645" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.531730] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.536162] env[61867]: DEBUG oslo_concurrency.lockutils [req-f5fe344d-bfda-4b67-a8e4-95a52c767ad0 req-867bc348-2459-4cbf-9f01-1bb6c0dede04 service nova] Acquired lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.536162] env[61867]: DEBUG nova.network.neutron [req-f5fe344d-bfda-4b67-a8e4-95a52c767ad0 req-867bc348-2459-4cbf-9f01-1bb6c0dede04 service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Refreshing network info cache for port c05702c4-5138-450d-82b2-790b86d8a59b {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1005.903387] env[61867]: DEBUG nova.scheduler.client.report [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1005.931507] env[61867]: DEBUG nova.objects.base [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Object Instance<0ae83268-ef50-42e6-954d-5e67c4afc20e> lazy-loaded attributes: flavor,pci_requests {{(pid=61867) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1005.931742] env[61867]: DEBUG nova.network.neutron [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1005.987960] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.013972] env[61867]: DEBUG nova.policy [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ea634246436422a9407f82692b69ede', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '86ced20eaf4740e298dc6f8ca5550c09', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 1006.024468] env[61867]: DEBUG nova.compute.manager [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1006.037277] env[61867]: DEBUG oslo_concurrency.lockutils [None req-e5113149-dc1d-482b-8f0b-91723be91f84 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "interface-4d7bfb31-d565-49e4-8c51-0122acebba2d-aa5b1005-fdcb-4208-b0d8-3b33573d05b6" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.862s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.375993] env[61867]: DEBUG nova.network.neutron [req-f5fe344d-bfda-4b67-a8e4-95a52c767ad0 req-867bc348-2459-4cbf-9f01-1bb6c0dede04 service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Updated VIF entry in instance network info cache for port c05702c4-5138-450d-82b2-790b86d8a59b. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1006.376396] env[61867]: DEBUG nova.network.neutron [req-f5fe344d-bfda-4b67-a8e4-95a52c767ad0 req-867bc348-2459-4cbf-9f01-1bb6c0dede04 service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Updating instance_info_cache with network_info: [{"id": "c05702c4-5138-450d-82b2-790b86d8a59b", "address": "fa:16:3e:cf:d4:2a", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc05702c4-51", "ovs_interfaceid": "c05702c4-5138-450d-82b2-790b86d8a59b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.408096] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.233s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.408665] env[61867]: DEBUG nova.compute.manager [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1006.411356] env[61867]: DEBUG oslo_concurrency.lockutils [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.779s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.412765] env[61867]: INFO nova.compute.claims [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1006.543743] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.879335] env[61867]: DEBUG oslo_concurrency.lockutils [req-f5fe344d-bfda-4b67-a8e4-95a52c767ad0 req-867bc348-2459-4cbf-9f01-1bb6c0dede04 service nova] Releasing lock "refresh_cache-4d7bfb31-d565-49e4-8c51-0122acebba2d" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.879619] env[61867]: DEBUG nova.compute.manager [req-f5fe344d-bfda-4b67-a8e4-95a52c767ad0 req-867bc348-2459-4cbf-9f01-1bb6c0dede04 service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Received event network-changed-a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.879799] env[61867]: DEBUG nova.compute.manager [req-f5fe344d-bfda-4b67-a8e4-95a52c767ad0 req-867bc348-2459-4cbf-9f01-1bb6c0dede04 service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Refreshing instance network info cache due to event network-changed-a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1006.880030] env[61867]: DEBUG oslo_concurrency.lockutils [req-f5fe344d-bfda-4b67-a8e4-95a52c767ad0 req-867bc348-2459-4cbf-9f01-1bb6c0dede04 service nova] Acquiring lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.880184] env[61867]: DEBUG oslo_concurrency.lockutils [req-f5fe344d-bfda-4b67-a8e4-95a52c767ad0 req-867bc348-2459-4cbf-9f01-1bb6c0dede04 service nova] Acquired lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.880409] env[61867]: DEBUG nova.network.neutron [req-f5fe344d-bfda-4b67-a8e4-95a52c767ad0 req-867bc348-2459-4cbf-9f01-1bb6c0dede04 service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Refreshing network info cache for port a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1006.917050] env[61867]: DEBUG nova.compute.utils [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1006.921057] env[61867]: DEBUG nova.compute.manager [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1006.921267] env[61867]: DEBUG nova.network.neutron [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1006.966988] env[61867]: DEBUG nova.policy [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '322fe9fc12a34183b4241e28f7254df4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24180ce8aca142fb897e29d4853c20c3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 1006.972888] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquiring lock "64947dbe-062c-4041-9ba9-d055279ded1e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.973138] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Lock "64947dbe-062c-4041-9ba9-d055279ded1e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.252840] env[61867]: DEBUG nova.network.neutron [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Successfully created port: eb146e5d-990b-477a-928e-ba9014b0fb40 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1007.422058] env[61867]: DEBUG nova.compute.manager [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1007.476439] env[61867]: DEBUG nova.compute.manager [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1007.588630] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14a9a339-8ae0-4318-b847-7aab0664534c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.600037] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b255ac6-5afb-4d50-8b6f-6bec17acb2ec {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.628704] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d63912-6c48-4376-a617-6b78aed1b038 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.638243] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3746d38-199a-4860-b7b3-a219c651fd99 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.651939] env[61867]: DEBUG nova.compute.provider_tree [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.761679] env[61867]: DEBUG nova.compute.manager [req-a21b8988-4c75-4e0e-9a03-8e87d2d9568a req-b5184f3b-4a80-4e08-a771-7d7708a35b1b service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Received event network-vif-plugged-aa5b1005-fdcb-4208-b0d8-3b33573d05b6 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.761917] env[61867]: DEBUG oslo_concurrency.lockutils [req-a21b8988-4c75-4e0e-9a03-8e87d2d9568a req-b5184f3b-4a80-4e08-a771-7d7708a35b1b service nova] Acquiring lock "0ae83268-ef50-42e6-954d-5e67c4afc20e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.762141] env[61867]: DEBUG oslo_concurrency.lockutils [req-a21b8988-4c75-4e0e-9a03-8e87d2d9568a req-b5184f3b-4a80-4e08-a771-7d7708a35b1b service nova] Lock "0ae83268-ef50-42e6-954d-5e67c4afc20e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.762308] env[61867]: DEBUG oslo_concurrency.lockutils [req-a21b8988-4c75-4e0e-9a03-8e87d2d9568a req-b5184f3b-4a80-4e08-a771-7d7708a35b1b service nova] Lock "0ae83268-ef50-42e6-954d-5e67c4afc20e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.762473] env[61867]: DEBUG nova.compute.manager [req-a21b8988-4c75-4e0e-9a03-8e87d2d9568a req-b5184f3b-4a80-4e08-a771-7d7708a35b1b service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] No waiting events found dispatching network-vif-plugged-aa5b1005-fdcb-4208-b0d8-3b33573d05b6 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1007.762669] env[61867]: WARNING nova.compute.manager [req-a21b8988-4c75-4e0e-9a03-8e87d2d9568a req-b5184f3b-4a80-4e08-a771-7d7708a35b1b service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Received unexpected event network-vif-plugged-aa5b1005-fdcb-4208-b0d8-3b33573d05b6 for instance with vm_state active and task_state None. [ 1007.801969] env[61867]: DEBUG nova.network.neutron [req-f5fe344d-bfda-4b67-a8e4-95a52c767ad0 req-867bc348-2459-4cbf-9f01-1bb6c0dede04 service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Updated VIF entry in instance network info cache for port a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1007.802746] env[61867]: DEBUG nova.network.neutron [req-f5fe344d-bfda-4b67-a8e4-95a52c767ad0 req-867bc348-2459-4cbf-9f01-1bb6c0dede04 service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Updating instance_info_cache with network_info: [{"id": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "address": "fa:16:3e:a0:23:06", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3537cac-b5", "ovs_interfaceid": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.874148] env[61867]: DEBUG nova.network.neutron [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Successfully updated port: aa5b1005-fdcb-4208-b0d8-3b33573d05b6 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1007.996772] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.154482] env[61867]: DEBUG nova.scheduler.client.report [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.305612] env[61867]: DEBUG oslo_concurrency.lockutils [req-f5fe344d-bfda-4b67-a8e4-95a52c767ad0 req-867bc348-2459-4cbf-9f01-1bb6c0dede04 service nova] Releasing lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.376672] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.376907] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.377199] env[61867]: DEBUG nova.network.neutron [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1008.434445] env[61867]: DEBUG nova.compute.manager [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1008.460957] env[61867]: DEBUG nova.virt.hardware [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1008.461241] env[61867]: DEBUG nova.virt.hardware [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1008.461404] env[61867]: DEBUG nova.virt.hardware [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1008.461587] env[61867]: DEBUG nova.virt.hardware [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1008.461737] env[61867]: DEBUG nova.virt.hardware [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1008.461888] env[61867]: DEBUG nova.virt.hardware [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1008.462108] env[61867]: DEBUG nova.virt.hardware [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1008.462272] env[61867]: DEBUG nova.virt.hardware [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1008.462441] env[61867]: DEBUG nova.virt.hardware [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1008.462606] env[61867]: DEBUG nova.virt.hardware [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1008.462846] env[61867]: DEBUG nova.virt.hardware [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1008.463720] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5adafa33-395a-4528-b382-1a6d1bbc74c0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.471401] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290ecab1-2c8b-41b7-8502-fca5ed5ed323 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.659409] env[61867]: DEBUG oslo_concurrency.lockutils [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.248s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.660065] env[61867]: DEBUG nova.compute.manager [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1008.663044] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.622s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.663131] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.665162] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.418s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.665361] env[61867]: DEBUG nova.objects.instance [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lazy-loading 'resources' on Instance uuid 6e41989e-b8fa-4009-af1e-1ce859b329a1 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.684202] env[61867]: INFO nova.scheduler.client.report [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleted allocations for instance 6e2136e7-c7e0-4a98-9899-f79c10f0e703 [ 1008.928554] env[61867]: WARNING nova.network.neutron [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] 7a62cd67-4e7d-4952-9a24-053e35c1c3cb already exists in list: networks containing: ['7a62cd67-4e7d-4952-9a24-053e35c1c3cb']. ignoring it [ 1008.938588] env[61867]: DEBUG nova.network.neutron [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Successfully updated port: eb146e5d-990b-477a-928e-ba9014b0fb40 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1009.168059] env[61867]: DEBUG nova.compute.utils [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1009.173169] env[61867]: DEBUG nova.compute.manager [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1009.173169] env[61867]: DEBUG nova.network.neutron [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1009.193633] env[61867]: DEBUG oslo_concurrency.lockutils [None req-bb3a2e28-0440-46c5-b68b-5d586352036b tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "6e2136e7-c7e0-4a98-9899-f79c10f0e703" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.037s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.204335] env[61867]: DEBUG nova.network.neutron [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Updating instance_info_cache with network_info: [{"id": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "address": "fa:16:3e:a0:23:06", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3537cac-b5", "ovs_interfaceid": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "aa5b1005-fdcb-4208-b0d8-3b33573d05b6", "address": "fa:16:3e:80:74:6e", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa5b1005-fd", "ovs_interfaceid": "aa5b1005-fdcb-4208-b0d8-3b33573d05b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.225609] env[61867]: DEBUG nova.policy [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12ed008ec3204102b2b08e61ed24f418', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '47c5b5b783ea445eb1c02fb728b1fff2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 1009.286922] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91e73e7-fa1e-4928-8072-0cb435abb50b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.294773] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-875c28e4-bfbd-446a-96a6-9a7454d7cb5f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.325446] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb602c6d-b4b8-4b94-857f-71ac57627fd6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.333125] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8de0051-c2c3-4bda-b529-effab53dc571 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.346247] env[61867]: DEBUG nova.compute.provider_tree [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.442367] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "refresh_cache-748e8a71-28b6-4c91-9cbb-856359348c96" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.442367] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquired lock "refresh_cache-748e8a71-28b6-4c91-9cbb-856359348c96" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.442367] env[61867]: DEBUG nova.network.neutron [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1009.466438] env[61867]: DEBUG nova.network.neutron [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Successfully created port: 380ad3aa-f8aa-41b7-94bb-71a4528022f2 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1009.672586] env[61867]: DEBUG nova.compute.manager [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1009.706630] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.707336] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.707508] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.708395] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3581e1b-61c0-4679-9ca8-6390260c6c75 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.725923] env[61867]: DEBUG nova.virt.hardware [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1009.726182] env[61867]: DEBUG nova.virt.hardware [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1009.726326] env[61867]: DEBUG nova.virt.hardware [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1009.726511] env[61867]: DEBUG nova.virt.hardware [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1009.726665] env[61867]: DEBUG nova.virt.hardware [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1009.726819] env[61867]: DEBUG nova.virt.hardware [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1009.727048] env[61867]: DEBUG nova.virt.hardware [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1009.727232] env[61867]: DEBUG nova.virt.hardware [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1009.727407] env[61867]: DEBUG nova.virt.hardware [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1009.727574] env[61867]: DEBUG nova.virt.hardware [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1009.727753] env[61867]: DEBUG nova.virt.hardware [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1009.733972] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Reconfiguring VM to attach interface {{(pid=61867) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1009.734506] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01f7f1dc-58df-4d62-8595-4c0c7726a7b4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.751979] env[61867]: DEBUG oslo_vmware.api [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 1009.751979] env[61867]: value = "task-1277058" [ 1009.751979] env[61867]: _type = "Task" [ 1009.751979] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.759773] env[61867]: DEBUG oslo_vmware.api [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277058, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.849609] env[61867]: DEBUG nova.scheduler.client.report [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1009.855203] env[61867]: DEBUG nova.compute.manager [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Received event network-changed-aa5b1005-fdcb-4208-b0d8-3b33573d05b6 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.855404] env[61867]: DEBUG nova.compute.manager [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Refreshing instance network info cache due to event network-changed-aa5b1005-fdcb-4208-b0d8-3b33573d05b6. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1009.855729] env[61867]: DEBUG oslo_concurrency.lockutils [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] Acquiring lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.855791] env[61867]: DEBUG oslo_concurrency.lockutils [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] Acquired lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.855928] env[61867]: DEBUG nova.network.neutron [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Refreshing network info cache for port aa5b1005-fdcb-4208-b0d8-3b33573d05b6 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1009.972400] env[61867]: DEBUG nova.network.neutron [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1010.102599] env[61867]: DEBUG nova.network.neutron [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Updating instance_info_cache with network_info: [{"id": "eb146e5d-990b-477a-928e-ba9014b0fb40", "address": "fa:16:3e:34:92:d2", "network": {"id": "fc403caf-421f-4783-9807-5e61c4379a6d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-846804042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24180ce8aca142fb897e29d4853c20c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb146e5d-99", "ovs_interfaceid": "eb146e5d-990b-477a-928e-ba9014b0fb40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.262062] env[61867]: DEBUG oslo_vmware.api [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277058, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.358339] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.693s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.362614] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 4.375s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.362854] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.363026] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61867) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1010.363327] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.820s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.364851] env[61867]: INFO nova.compute.claims [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1010.369522] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bf9f892-a282-4990-a24c-f1b505ea978a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.378450] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-287a3d47-d5e0-4f14-b9e7-f2b9a3424070 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.384206] env[61867]: INFO nova.scheduler.client.report [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Deleted allocations for instance 6e41989e-b8fa-4009-af1e-1ce859b329a1 [ 1010.396497] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56fd504-18fa-42d6-8de5-e2a574e23cb0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.403498] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-717cf240-f463-4a3a-998d-b1dd9176e9ef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.437171] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180286MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61867) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1010.437331] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.574796] env[61867]: DEBUG nova.network.neutron [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Updated VIF entry in instance network info cache for port aa5b1005-fdcb-4208-b0d8-3b33573d05b6. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1010.575354] env[61867]: DEBUG nova.network.neutron [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Updating instance_info_cache with network_info: [{"id": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "address": "fa:16:3e:a0:23:06", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3537cac-b5", "ovs_interfaceid": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "aa5b1005-fdcb-4208-b0d8-3b33573d05b6", "address": "fa:16:3e:80:74:6e", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa5b1005-fd", "ovs_interfaceid": "aa5b1005-fdcb-4208-b0d8-3b33573d05b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.605209] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Releasing lock "refresh_cache-748e8a71-28b6-4c91-9cbb-856359348c96" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.605543] env[61867]: DEBUG nova.compute.manager [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Instance network_info: |[{"id": "eb146e5d-990b-477a-928e-ba9014b0fb40", "address": "fa:16:3e:34:92:d2", "network": {"id": "fc403caf-421f-4783-9807-5e61c4379a6d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-846804042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24180ce8aca142fb897e29d4853c20c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb146e5d-99", "ovs_interfaceid": "eb146e5d-990b-477a-928e-ba9014b0fb40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1010.605980] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:34:92:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '47ca1ce6-8148-48d5-bcfe-89e39b73914e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eb146e5d-990b-477a-928e-ba9014b0fb40', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1010.613417] env[61867]: DEBUG oslo.service.loopingcall [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1010.613624] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1010.613887] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a96e331b-61b1-43fe-80ef-afea70e175f5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.635153] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1010.635153] env[61867]: value = "task-1277059" [ 1010.635153] env[61867]: _type = "Task" [ 1010.635153] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.643026] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277059, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.683214] env[61867]: DEBUG nova.compute.manager [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1010.710991] env[61867]: DEBUG nova.virt.hardware [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1010.711279] env[61867]: DEBUG nova.virt.hardware [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1010.711439] env[61867]: DEBUG nova.virt.hardware [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1010.711627] env[61867]: DEBUG nova.virt.hardware [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1010.711775] env[61867]: DEBUG nova.virt.hardware [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1010.711927] env[61867]: DEBUG nova.virt.hardware [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1010.712198] env[61867]: DEBUG nova.virt.hardware [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1010.712381] env[61867]: DEBUG nova.virt.hardware [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1010.712554] env[61867]: DEBUG nova.virt.hardware [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1010.712754] env[61867]: DEBUG nova.virt.hardware [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1010.712941] env[61867]: DEBUG nova.virt.hardware [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1010.713836] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db2c7ee-bdee-4d6d-b4f7-b2e067533edd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.722039] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01607180-61d8-4080-ba17-8f4ac2cbb3a2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.761238] env[61867]: DEBUG oslo_vmware.api [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277058, 'name': ReconfigVM_Task, 'duration_secs': 0.519553} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.761718] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.761943] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Reconfigured VM to attach interface {{(pid=61867) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1010.824554] env[61867]: DEBUG nova.compute.manager [req-2a3f3872-cdd0-452b-b566-0409a6bae78d req-9e5d6f85-39ea-4185-9956-fb81ed9a6aac service nova] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Received event network-vif-plugged-380ad3aa-f8aa-41b7-94bb-71a4528022f2 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1010.824765] env[61867]: DEBUG oslo_concurrency.lockutils [req-2a3f3872-cdd0-452b-b566-0409a6bae78d req-9e5d6f85-39ea-4185-9956-fb81ed9a6aac service nova] Acquiring lock "29d961df-5db5-46db-ace9-651913e0250a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.824990] env[61867]: DEBUG oslo_concurrency.lockutils [req-2a3f3872-cdd0-452b-b566-0409a6bae78d req-9e5d6f85-39ea-4185-9956-fb81ed9a6aac service nova] Lock "29d961df-5db5-46db-ace9-651913e0250a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.825171] env[61867]: DEBUG oslo_concurrency.lockutils [req-2a3f3872-cdd0-452b-b566-0409a6bae78d req-9e5d6f85-39ea-4185-9956-fb81ed9a6aac service nova] Lock "29d961df-5db5-46db-ace9-651913e0250a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.825347] env[61867]: DEBUG nova.compute.manager [req-2a3f3872-cdd0-452b-b566-0409a6bae78d req-9e5d6f85-39ea-4185-9956-fb81ed9a6aac service nova] [instance: 29d961df-5db5-46db-ace9-651913e0250a] No waiting events found dispatching network-vif-plugged-380ad3aa-f8aa-41b7-94bb-71a4528022f2 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1010.825511] env[61867]: WARNING nova.compute.manager [req-2a3f3872-cdd0-452b-b566-0409a6bae78d req-9e5d6f85-39ea-4185-9956-fb81ed9a6aac service nova] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Received unexpected event network-vif-plugged-380ad3aa-f8aa-41b7-94bb-71a4528022f2 for instance with vm_state building and task_state spawning. [ 1010.898633] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9deed3aa-7f86-4a9b-8b86-8d77b037e774 tempest-ServerRescueNegativeTestJSON-1513557601 tempest-ServerRescueNegativeTestJSON-1513557601-project-member] Lock "6e41989e-b8fa-4009-af1e-1ce859b329a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.230s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.078259] env[61867]: DEBUG oslo_concurrency.lockutils [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] Releasing lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.078561] env[61867]: DEBUG nova.compute.manager [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Received event network-vif-plugged-eb146e5d-990b-477a-928e-ba9014b0fb40 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.078764] env[61867]: DEBUG oslo_concurrency.lockutils [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] Acquiring lock "748e8a71-28b6-4c91-9cbb-856359348c96-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.078978] env[61867]: DEBUG oslo_concurrency.lockutils [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] Lock "748e8a71-28b6-4c91-9cbb-856359348c96-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.079164] env[61867]: DEBUG oslo_concurrency.lockutils [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] Lock "748e8a71-28b6-4c91-9cbb-856359348c96-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.079332] env[61867]: DEBUG nova.compute.manager [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] No waiting events found dispatching network-vif-plugged-eb146e5d-990b-477a-928e-ba9014b0fb40 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1011.079501] env[61867]: WARNING nova.compute.manager [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Received unexpected event network-vif-plugged-eb146e5d-990b-477a-928e-ba9014b0fb40 for instance with vm_state building and task_state spawning. [ 1011.079664] env[61867]: DEBUG nova.compute.manager [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Received event network-changed-eb146e5d-990b-477a-928e-ba9014b0fb40 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.079905] env[61867]: DEBUG nova.compute.manager [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Refreshing instance network info cache due to event network-changed-eb146e5d-990b-477a-928e-ba9014b0fb40. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1011.080138] env[61867]: DEBUG oslo_concurrency.lockutils [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] Acquiring lock "refresh_cache-748e8a71-28b6-4c91-9cbb-856359348c96" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.080284] env[61867]: DEBUG oslo_concurrency.lockutils [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] Acquired lock "refresh_cache-748e8a71-28b6-4c91-9cbb-856359348c96" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.080448] env[61867]: DEBUG nova.network.neutron [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Refreshing network info cache for port eb146e5d-990b-477a-928e-ba9014b0fb40 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1011.145097] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277059, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.266551] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a2974d29-e382-4b10-879d-4f097e805355 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "interface-0ae83268-ef50-42e6-954d-5e67c4afc20e-aa5b1005-fdcb-4208-b0d8-3b33573d05b6" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.457s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.412772] env[61867]: DEBUG nova.network.neutron [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Successfully updated port: 380ad3aa-f8aa-41b7-94bb-71a4528022f2 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1011.489816] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff24123a-c6d3-41fe-9cf1-8b4c4a25572b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.497919] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e2121f-9eda-431a-bc0e-dd56a83a0810 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.528794] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f822d3a6-ed6f-4420-929c-82911039fada {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.536578] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a1c36d-5254-439b-bfcd-e39d4952cf0a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.551420] env[61867]: DEBUG nova.compute.provider_tree [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.645818] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277059, 'name': CreateVM_Task, 'duration_secs': 0.823664} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.646141] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1011.646727] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.646944] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.647319] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1011.647605] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f52e9c3f-340d-4444-a7d4-8128e77fe457 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.651975] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1011.651975] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ca2c0e-8d03-d98b-de0f-b2dfc213a030" [ 1011.651975] env[61867]: _type = "Task" [ 1011.651975] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.659124] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ca2c0e-8d03-d98b-de0f-b2dfc213a030, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.877759] env[61867]: DEBUG nova.compute.manager [req-aaef8e9e-a906-4a8e-889c-769f472278ec req-a4ddb4cd-7e24-4e1a-bf47-2afd466f87ee service nova] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Received event network-changed-380ad3aa-f8aa-41b7-94bb-71a4528022f2 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.877966] env[61867]: DEBUG nova.compute.manager [req-aaef8e9e-a906-4a8e-889c-769f472278ec req-a4ddb4cd-7e24-4e1a-bf47-2afd466f87ee service nova] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Refreshing instance network info cache due to event network-changed-380ad3aa-f8aa-41b7-94bb-71a4528022f2. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1011.878204] env[61867]: DEBUG oslo_concurrency.lockutils [req-aaef8e9e-a906-4a8e-889c-769f472278ec req-a4ddb4cd-7e24-4e1a-bf47-2afd466f87ee service nova] Acquiring lock "refresh_cache-29d961df-5db5-46db-ace9-651913e0250a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.878356] env[61867]: DEBUG oslo_concurrency.lockutils [req-aaef8e9e-a906-4a8e-889c-769f472278ec req-a4ddb4cd-7e24-4e1a-bf47-2afd466f87ee service nova] Acquired lock "refresh_cache-29d961df-5db5-46db-ace9-651913e0250a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.878519] env[61867]: DEBUG nova.network.neutron [req-aaef8e9e-a906-4a8e-889c-769f472278ec req-a4ddb4cd-7e24-4e1a-bf47-2afd466f87ee service nova] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Refreshing network info cache for port 380ad3aa-f8aa-41b7-94bb-71a4528022f2 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1011.916310] env[61867]: DEBUG oslo_concurrency.lockutils [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "refresh_cache-29d961df-5db5-46db-ace9-651913e0250a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.054870] env[61867]: DEBUG nova.scheduler.client.report [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.102584] env[61867]: DEBUG nova.network.neutron [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Updated VIF entry in instance network info cache for port eb146e5d-990b-477a-928e-ba9014b0fb40. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1012.103203] env[61867]: DEBUG nova.network.neutron [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Updating instance_info_cache with network_info: [{"id": "eb146e5d-990b-477a-928e-ba9014b0fb40", "address": "fa:16:3e:34:92:d2", "network": {"id": "fc403caf-421f-4783-9807-5e61c4379a6d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-846804042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24180ce8aca142fb897e29d4853c20c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb146e5d-99", "ovs_interfaceid": "eb146e5d-990b-477a-928e-ba9014b0fb40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.162469] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ca2c0e-8d03-d98b-de0f-b2dfc213a030, 'name': SearchDatastore_Task, 'duration_secs': 0.009947} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.162987] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.163064] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1012.163311] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.163466] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.163692] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1012.163944] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1edcf9d1-8baa-4ea6-ae38-3c6c3060a8ba {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.172197] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1012.172408] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1012.173179] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32b68378-6087-4fc6-97b2-46e4a022eb23 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.178118] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1012.178118] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5264c674-5852-aec8-f06c-b31afd050579" [ 1012.178118] env[61867]: _type = "Task" [ 1012.178118] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.186722] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5264c674-5852-aec8-f06c-b31afd050579, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.428739] env[61867]: DEBUG nova.network.neutron [req-aaef8e9e-a906-4a8e-889c-769f472278ec req-a4ddb4cd-7e24-4e1a-bf47-2afd466f87ee service nova] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1012.527220] env[61867]: DEBUG oslo_concurrency.lockutils [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "interface-0ae83268-ef50-42e6-954d-5e67c4afc20e-aa5b1005-fdcb-4208-b0d8-3b33573d05b6" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.527591] env[61867]: DEBUG oslo_concurrency.lockutils [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "interface-0ae83268-ef50-42e6-954d-5e67c4afc20e-aa5b1005-fdcb-4208-b0d8-3b33573d05b6" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.544282] env[61867]: DEBUG nova.network.neutron [req-aaef8e9e-a906-4a8e-889c-769f472278ec req-a4ddb4cd-7e24-4e1a-bf47-2afd466f87ee service nova] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.559323] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.196s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.559832] env[61867]: DEBUG nova.compute.manager [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1012.566289] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.566s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.567714] env[61867]: INFO nova.compute.claims [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1012.605990] env[61867]: DEBUG oslo_concurrency.lockutils [req-50b05999-2621-4b5e-b35a-8a200d0b4f34 req-0bac4bbc-351b-47d8-8232-6eb14dc88ca2 service nova] Releasing lock "refresh_cache-748e8a71-28b6-4c91-9cbb-856359348c96" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.689630] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5264c674-5852-aec8-f06c-b31afd050579, 'name': SearchDatastore_Task, 'duration_secs': 0.008553} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.690518] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82ec17c2-6d77-4736-8d24-92548526e56b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.696112] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1012.696112] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529e7796-a307-c7f4-eac5-c63f6529d659" [ 1012.696112] env[61867]: _type = "Task" [ 1012.696112] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.704268] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529e7796-a307-c7f4-eac5-c63f6529d659, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.031737] env[61867]: DEBUG oslo_concurrency.lockutils [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.031952] env[61867]: DEBUG oslo_concurrency.lockutils [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.032860] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0032e09-368e-4008-a907-8236e120ba95 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.050317] env[61867]: DEBUG oslo_concurrency.lockutils [req-aaef8e9e-a906-4a8e-889c-769f472278ec req-a4ddb4cd-7e24-4e1a-bf47-2afd466f87ee service nova] Releasing lock "refresh_cache-29d961df-5db5-46db-ace9-651913e0250a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.051104] env[61867]: DEBUG oslo_concurrency.lockutils [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "refresh_cache-29d961df-5db5-46db-ace9-651913e0250a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.051104] env[61867]: DEBUG nova.network.neutron [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1013.052739] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b87953a9-5c55-4d7d-b3d7-0ddc1772e39a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.074269] env[61867]: DEBUG nova.compute.utils [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1013.082836] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Reconfiguring VM to detach interface {{(pid=61867) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1013.084216] env[61867]: DEBUG nova.compute.manager [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1013.084572] env[61867]: DEBUG nova.network.neutron [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1013.086190] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d47ef43b-fe52-4081-b5f0-a6ab1e99d231 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.106307] env[61867]: DEBUG oslo_vmware.api [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 1013.106307] env[61867]: value = "task-1277060" [ 1013.106307] env[61867]: _type = "Task" [ 1013.106307] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.116515] env[61867]: DEBUG oslo_vmware.api [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277060, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.145601] env[61867]: DEBUG nova.policy [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a013ca682054d16a6702230a22c180f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c28a8e133444440eb83dc6a848ef591d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 1013.206247] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529e7796-a307-c7f4-eac5-c63f6529d659, 'name': SearchDatastore_Task, 'duration_secs': 0.008937} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.206536] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.206795] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 748e8a71-28b6-4c91-9cbb-856359348c96/748e8a71-28b6-4c91-9cbb-856359348c96.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1013.207064] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0bd120f9-fc28-411f-bf38-547e0273a780 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.214214] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1013.214214] env[61867]: value = "task-1277061" [ 1013.214214] env[61867]: _type = "Task" [ 1013.214214] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.221753] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277061, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.462376] env[61867]: DEBUG nova.network.neutron [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Successfully created port: 072ec5db-d2ce-4e94-809d-7ff5fd034562 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1013.589068] env[61867]: DEBUG nova.compute.manager [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1013.602462] env[61867]: DEBUG nova.network.neutron [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1013.620726] env[61867]: DEBUG oslo_vmware.api [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.700582] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec1c25d-18dc-4bbd-bcf6-ee15877e51de {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.709731] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bb64e94-ce16-4038-9303-3b3092fa6284 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.741452] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7440c1bd-4bbe-45a4-9645-c8a7e4923a29 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.746439] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277061, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.441094} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.746998] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 748e8a71-28b6-4c91-9cbb-856359348c96/748e8a71-28b6-4c91-9cbb-856359348c96.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1013.747247] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1013.747471] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1858dc76-db29-4c40-be20-7d5505ea80dd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.752202] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5c83f3-ca74-44a8-8da3-2c06d17f6c25 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.756623] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1013.756623] env[61867]: value = "task-1277062" [ 1013.756623] env[61867]: _type = "Task" [ 1013.756623] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.769737] env[61867]: DEBUG nova.compute.provider_tree [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.775547] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277062, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.823839] env[61867]: DEBUG nova.network.neutron [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Updating instance_info_cache with network_info: [{"id": "380ad3aa-f8aa-41b7-94bb-71a4528022f2", "address": "fa:16:3e:ca:a2:9f", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap380ad3aa-f8", "ovs_interfaceid": "380ad3aa-f8aa-41b7-94bb-71a4528022f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.116951] env[61867]: DEBUG oslo_vmware.api [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.268306] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277062, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060637} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.268573] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1014.269379] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1778d43-4619-476c-9816-1203c48791d9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.272371] env[61867]: DEBUG nova.scheduler.client.report [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.294721] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 748e8a71-28b6-4c91-9cbb-856359348c96/748e8a71-28b6-4c91-9cbb-856359348c96.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1014.295604] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f57d87fb-7f3b-4c7f-a762-2ec4b4468ce1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.314308] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1014.314308] env[61867]: value = "task-1277063" [ 1014.314308] env[61867]: _type = "Task" [ 1014.314308] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.322048] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277063, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.326782] env[61867]: DEBUG oslo_concurrency.lockutils [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "refresh_cache-29d961df-5db5-46db-ace9-651913e0250a" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.327091] env[61867]: DEBUG nova.compute.manager [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Instance network_info: |[{"id": "380ad3aa-f8aa-41b7-94bb-71a4528022f2", "address": "fa:16:3e:ca:a2:9f", "network": {"id": "c1d1854f-0b2b-4c37-815a-7c48255797f4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-555057390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47c5b5b783ea445eb1c02fb728b1fff2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2020f39-42c4-4481-85c5-aaf03854b459", "external-id": "nsx-vlan-transportzone-802", "segmentation_id": 802, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap380ad3aa-f8", "ovs_interfaceid": "380ad3aa-f8aa-41b7-94bb-71a4528022f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1014.327470] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:a2:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f2020f39-42c4-4481-85c5-aaf03854b459', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '380ad3aa-f8aa-41b7-94bb-71a4528022f2', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1014.335033] env[61867]: DEBUG oslo.service.loopingcall [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.335243] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1014.335705] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3481dd19-3eb8-46ad-998b-b7f0f63a2f17 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.354458] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1014.354458] env[61867]: value = "task-1277064" [ 1014.354458] env[61867]: _type = "Task" [ 1014.354458] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.364254] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277064, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.599872] env[61867]: DEBUG nova.compute.manager [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1014.618838] env[61867]: DEBUG oslo_vmware.api [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.626908] env[61867]: DEBUG nova.virt.hardware [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1014.627083] env[61867]: DEBUG nova.virt.hardware [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1014.627247] env[61867]: DEBUG nova.virt.hardware [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1014.627433] env[61867]: DEBUG nova.virt.hardware [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1014.627581] env[61867]: DEBUG nova.virt.hardware [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1014.627732] env[61867]: DEBUG nova.virt.hardware [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1014.628156] env[61867]: DEBUG nova.virt.hardware [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1014.628156] env[61867]: DEBUG nova.virt.hardware [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1014.628416] env[61867]: DEBUG nova.virt.hardware [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1014.628464] env[61867]: DEBUG nova.virt.hardware [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1014.628613] env[61867]: DEBUG nova.virt.hardware [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1014.629415] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a47e9f-36ba-423c-ad92-f130d2065bdb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.636641] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e985f3b-df45-4a16-a479-9962f7592b45 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.777599] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.215s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.778166] env[61867]: DEBUG nova.compute.manager [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1014.780718] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.343s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.824152] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277063, 'name': ReconfigVM_Task, 'duration_secs': 0.391624} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.824420] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 748e8a71-28b6-4c91-9cbb-856359348c96/748e8a71-28b6-4c91-9cbb-856359348c96.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.825019] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0bc69a98-93d0-40bf-a3ba-4282cff48810 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.830966] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1014.830966] env[61867]: value = "task-1277065" [ 1014.830966] env[61867]: _type = "Task" [ 1014.830966] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.838799] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277065, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.863585] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277064, 'name': CreateVM_Task, 'duration_secs': 0.366133} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.863693] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1014.864377] env[61867]: DEBUG oslo_concurrency.lockutils [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.864684] env[61867]: DEBUG oslo_concurrency.lockutils [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.864944] env[61867]: DEBUG oslo_concurrency.lockutils [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1014.865224] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eeb7d3da-db42-4dc5-8ad1-38cb4f9b3e7f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.869480] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 1014.869480] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5227cac2-cafc-ab3d-8bc3-6edc1e0828b9" [ 1014.869480] env[61867]: _type = "Task" [ 1014.869480] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.876864] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5227cac2-cafc-ab3d-8bc3-6edc1e0828b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.984815] env[61867]: DEBUG nova.compute.manager [req-07095bc2-dcef-4683-8c3f-d20c0a06723c req-746e5f20-4dd2-418c-acd9-1ae300c5f6bb service nova] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Received event network-vif-plugged-072ec5db-d2ce-4e94-809d-7ff5fd034562 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.985081] env[61867]: DEBUG oslo_concurrency.lockutils [req-07095bc2-dcef-4683-8c3f-d20c0a06723c req-746e5f20-4dd2-418c-acd9-1ae300c5f6bb service nova] Acquiring lock "986ada58-eb26-4289-bc0e-338c6a399645-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.985309] env[61867]: DEBUG oslo_concurrency.lockutils [req-07095bc2-dcef-4683-8c3f-d20c0a06723c req-746e5f20-4dd2-418c-acd9-1ae300c5f6bb service nova] Lock "986ada58-eb26-4289-bc0e-338c6a399645-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.985483] env[61867]: DEBUG oslo_concurrency.lockutils [req-07095bc2-dcef-4683-8c3f-d20c0a06723c req-746e5f20-4dd2-418c-acd9-1ae300c5f6bb service nova] Lock "986ada58-eb26-4289-bc0e-338c6a399645-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.985734] env[61867]: DEBUG nova.compute.manager [req-07095bc2-dcef-4683-8c3f-d20c0a06723c req-746e5f20-4dd2-418c-acd9-1ae300c5f6bb service nova] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] No waiting events found dispatching network-vif-plugged-072ec5db-d2ce-4e94-809d-7ff5fd034562 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1014.985946] env[61867]: WARNING nova.compute.manager [req-07095bc2-dcef-4683-8c3f-d20c0a06723c req-746e5f20-4dd2-418c-acd9-1ae300c5f6bb service nova] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Received unexpected event network-vif-plugged-072ec5db-d2ce-4e94-809d-7ff5fd034562 for instance with vm_state building and task_state spawning. [ 1015.078822] env[61867]: DEBUG nova.network.neutron [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Successfully updated port: 072ec5db-d2ce-4e94-809d-7ff5fd034562 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1015.117971] env[61867]: DEBUG oslo_vmware.api [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.284477] env[61867]: DEBUG nova.compute.utils [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1015.285896] env[61867]: DEBUG nova.compute.manager [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Not allocating networking since 'none' was specified. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1015.340469] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277065, 'name': Rename_Task, 'duration_secs': 0.134763} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.340827] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1015.340982] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ced31e1-fc31-4854-8e4c-6c6f6553a74e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.347539] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1015.347539] env[61867]: value = "task-1277066" [ 1015.347539] env[61867]: _type = "Task" [ 1015.347539] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.354835] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277066, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.379110] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5227cac2-cafc-ab3d-8bc3-6edc1e0828b9, 'name': SearchDatastore_Task, 'duration_secs': 0.009131} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.379405] env[61867]: DEBUG oslo_concurrency.lockutils [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.379659] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1015.379913] env[61867]: DEBUG oslo_concurrency.lockutils [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.380949] env[61867]: DEBUG oslo_concurrency.lockutils [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.380949] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1015.380949] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1cb9937f-9600-43e9-a57e-96521162854a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.389220] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1015.389478] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1015.390175] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e927412b-f8f5-4819-af30-a6601cc66646 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.395041] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 1015.395041] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525dc14e-7dbf-a89a-fa7f-fc72fcd7f304" [ 1015.395041] env[61867]: _type = "Task" [ 1015.395041] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.401984] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525dc14e-7dbf-a89a-fa7f-fc72fcd7f304, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.581322] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "refresh_cache-986ada58-eb26-4289-bc0e-338c6a399645" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.581478] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "refresh_cache-986ada58-eb26-4289-bc0e-338c6a399645" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.581632] env[61867]: DEBUG nova.network.neutron [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1015.618087] env[61867]: DEBUG oslo_vmware.api [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.794693] env[61867]: DEBUG nova.compute.manager [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1015.813769] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 4d7bfb31-d565-49e4-8c51-0122acebba2d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.813945] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 0ae83268-ef50-42e6-954d-5e67c4afc20e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.814087] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.814211] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 748e8a71-28b6-4c91-9cbb-856359348c96 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.814330] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 29d961df-5db5-46db-ace9-651913e0250a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.814468] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 986ada58-eb26-4289-bc0e-338c6a399645 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.814589] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 64947dbe-062c-4041-9ba9-d055279ded1e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.814776] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1015.814919] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1015.858550] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277066, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.900313] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91b2a3fb-7e44-43b5-b17e-9247f8e01256 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.911061] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da650fab-b7b5-43da-a701-a81c9490daef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.913991] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525dc14e-7dbf-a89a-fa7f-fc72fcd7f304, 'name': SearchDatastore_Task, 'duration_secs': 0.007941} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.915044] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-184ff316-3064-4169-a142-f0e43d853259 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.944416] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93d715a0-fe6c-40dc-b96e-7e2cd8c270e3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.946989] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 1015.946989] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52de996f-5d07-cb6f-87e2-1e0b0774b62b" [ 1015.946989] env[61867]: _type = "Task" [ 1015.946989] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.953134] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-257f1c7c-ab6a-4b84-8bb2-bf4937dfe8e9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.959937] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52de996f-5d07-cb6f-87e2-1e0b0774b62b, 'name': SearchDatastore_Task, 'duration_secs': 0.009032} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.960230] env[61867]: DEBUG oslo_concurrency.lockutils [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.960484] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 29d961df-5db5-46db-ace9-651913e0250a/29d961df-5db5-46db-ace9-651913e0250a.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1015.960721] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c79b2d47-8976-44cb-9181-8b1b48fc6ead {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.969824] env[61867]: DEBUG nova.compute.provider_tree [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.974791] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 1015.974791] env[61867]: value = "task-1277067" [ 1015.974791] env[61867]: _type = "Task" [ 1015.974791] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.982953] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277067, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.118820] env[61867]: DEBUG oslo_vmware.api [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.121334] env[61867]: DEBUG nova.network.neutron [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1016.276522] env[61867]: DEBUG nova.network.neutron [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Updating instance_info_cache with network_info: [{"id": "072ec5db-d2ce-4e94-809d-7ff5fd034562", "address": "fa:16:3e:45:de:03", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap072ec5db-d2", "ovs_interfaceid": "072ec5db-d2ce-4e94-809d-7ff5fd034562", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.359247] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277066, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.473475] env[61867]: DEBUG nova.scheduler.client.report [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1016.487052] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277067, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.436191} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.487361] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 29d961df-5db5-46db-ace9-651913e0250a/29d961df-5db5-46db-ace9-651913e0250a.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1016.487579] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1016.487831] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-41b58606-b579-4777-84b3-9a2f93a30471 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.495470] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 1016.495470] env[61867]: value = "task-1277068" [ 1016.495470] env[61867]: _type = "Task" [ 1016.495470] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.505125] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277068, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.619206] env[61867]: DEBUG oslo_vmware.api [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.779123] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "refresh_cache-986ada58-eb26-4289-bc0e-338c6a399645" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.779477] env[61867]: DEBUG nova.compute.manager [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Instance network_info: |[{"id": "072ec5db-d2ce-4e94-809d-7ff5fd034562", "address": "fa:16:3e:45:de:03", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap072ec5db-d2", "ovs_interfaceid": "072ec5db-d2ce-4e94-809d-7ff5fd034562", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1016.779921] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:de:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74e6f6e0-95e6-4531-99e9-0e78350fb655', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '072ec5db-d2ce-4e94-809d-7ff5fd034562', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1016.787963] env[61867]: DEBUG oslo.service.loopingcall [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.788205] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1016.788433] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8730dd9b-c5b5-4c86-ac7f-d693528fd538 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.803686] env[61867]: DEBUG nova.compute.manager [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1016.813151] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1016.813151] env[61867]: value = "task-1277069" [ 1016.813151] env[61867]: _type = "Task" [ 1016.813151] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.821297] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277069, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.824314] env[61867]: DEBUG nova.virt.hardware [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1016.824556] env[61867]: DEBUG nova.virt.hardware [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1016.824716] env[61867]: DEBUG nova.virt.hardware [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1016.824903] env[61867]: DEBUG nova.virt.hardware [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1016.825072] env[61867]: DEBUG nova.virt.hardware [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1016.825257] env[61867]: DEBUG nova.virt.hardware [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1016.825536] env[61867]: DEBUG nova.virt.hardware [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1016.825756] env[61867]: DEBUG nova.virt.hardware [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1016.825950] env[61867]: DEBUG nova.virt.hardware [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1016.826145] env[61867]: DEBUG nova.virt.hardware [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1016.826329] env[61867]: DEBUG nova.virt.hardware [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1016.827359] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8115f6-8607-4d16-bc7b-dbbd65f541d0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.835218] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50b0c76-df6e-47ee-81c0-52e3ca5ce2bf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.848638] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Instance VIF info [] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1016.854199] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Creating folder: Project (4798405712b7464c9d157375666f09ce). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1016.854539] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-330def33-b9ee-4434-9beb-e3cd352fdb0a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.865751] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277066, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.867710] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Created folder: Project (4798405712b7464c9d157375666f09ce) in parent group-v274258. [ 1016.867921] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Creating folder: Instances. Parent ref: group-v274416. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1016.868233] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-643a99be-62e6-4c27-b121-94aa3331b32e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.877041] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Created folder: Instances in parent group-v274416. [ 1016.877300] env[61867]: DEBUG oslo.service.loopingcall [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.877508] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1016.877728] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2e14a345-56fd-4330-9760-d9f5b7b1f777 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.895455] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1016.895455] env[61867]: value = "task-1277072" [ 1016.895455] env[61867]: _type = "Task" [ 1016.895455] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.903898] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277072, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.978123] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61867) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1016.978348] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.198s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.005512] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277068, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.169188} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.005799] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1017.006617] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a9b9d8-e718-477d-bc5a-e0fb996e91ef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.011451] env[61867]: DEBUG nova.compute.manager [req-d127d5b5-d0b0-4534-8939-b16e627c9564 req-4eda00b4-f88c-46f3-b812-d58bbe8a2c2e service nova] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Received event network-changed-072ec5db-d2ce-4e94-809d-7ff5fd034562 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1017.011595] env[61867]: DEBUG nova.compute.manager [req-d127d5b5-d0b0-4534-8939-b16e627c9564 req-4eda00b4-f88c-46f3-b812-d58bbe8a2c2e service nova] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Refreshing instance network info cache due to event network-changed-072ec5db-d2ce-4e94-809d-7ff5fd034562. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1017.011816] env[61867]: DEBUG oslo_concurrency.lockutils [req-d127d5b5-d0b0-4534-8939-b16e627c9564 req-4eda00b4-f88c-46f3-b812-d58bbe8a2c2e service nova] Acquiring lock "refresh_cache-986ada58-eb26-4289-bc0e-338c6a399645" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.011964] env[61867]: DEBUG oslo_concurrency.lockutils [req-d127d5b5-d0b0-4534-8939-b16e627c9564 req-4eda00b4-f88c-46f3-b812-d58bbe8a2c2e service nova] Acquired lock "refresh_cache-986ada58-eb26-4289-bc0e-338c6a399645" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.012144] env[61867]: DEBUG nova.network.neutron [req-d127d5b5-d0b0-4534-8939-b16e627c9564 req-4eda00b4-f88c-46f3-b812-d58bbe8a2c2e service nova] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Refreshing network info cache for port 072ec5db-d2ce-4e94-809d-7ff5fd034562 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1017.034219] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 29d961df-5db5-46db-ace9-651913e0250a/29d961df-5db5-46db-ace9-651913e0250a.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1017.034986] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6bcf4d1d-46a9-4737-8528-98fb08fd1123 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.054919] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 1017.054919] env[61867]: value = "task-1277073" [ 1017.054919] env[61867]: _type = "Task" [ 1017.054919] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.063734] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277073, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.121041] env[61867]: DEBUG oslo_vmware.api [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.322888] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277069, 'name': CreateVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.364075] env[61867]: DEBUG oslo_vmware.api [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277066, 'name': PowerOnVM_Task, 'duration_secs': 1.591887} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.364422] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1017.364652] env[61867]: INFO nova.compute.manager [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Took 8.93 seconds to spawn the instance on the hypervisor. [ 1017.364881] env[61867]: DEBUG nova.compute.manager [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1017.365698] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db6c7f7-c7ba-4f1b-b5b6-8c6e3eacb71d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.405796] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277072, 'name': CreateVM_Task, 'duration_secs': 0.481934} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.405925] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1017.406353] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.406519] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.406903] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1017.407830] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d764e66-ee28-4ded-8eb7-a301983adff6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.412483] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1017.412483] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520780d4-d43e-dad5-90d1-b0ef592ee0a6" [ 1017.412483] env[61867]: _type = "Task" [ 1017.412483] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.420133] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520780d4-d43e-dad5-90d1-b0ef592ee0a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.565051] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277073, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.619509] env[61867]: DEBUG oslo_vmware.api [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.715531] env[61867]: DEBUG nova.network.neutron [req-d127d5b5-d0b0-4534-8939-b16e627c9564 req-4eda00b4-f88c-46f3-b812-d58bbe8a2c2e service nova] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Updated VIF entry in instance network info cache for port 072ec5db-d2ce-4e94-809d-7ff5fd034562. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1017.715907] env[61867]: DEBUG nova.network.neutron [req-d127d5b5-d0b0-4534-8939-b16e627c9564 req-4eda00b4-f88c-46f3-b812-d58bbe8a2c2e service nova] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Updating instance_info_cache with network_info: [{"id": "072ec5db-d2ce-4e94-809d-7ff5fd034562", "address": "fa:16:3e:45:de:03", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap072ec5db-d2", "ovs_interfaceid": "072ec5db-d2ce-4e94-809d-7ff5fd034562", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.823789] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277069, 'name': CreateVM_Task, 'duration_secs': 0.551112} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.824124] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1017.824593] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.886047] env[61867]: INFO nova.compute.manager [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Took 17.58 seconds to build instance. [ 1017.923146] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520780d4-d43e-dad5-90d1-b0ef592ee0a6, 'name': SearchDatastore_Task, 'duration_secs': 0.011066} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.923527] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.923745] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1017.924121] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.924198] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.924385] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1017.924711] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.924974] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1017.925237] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a41105d6-c1d9-4796-861d-ec991fa4a43d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.927170] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e68a7d4b-4b2a-4d49-b56c-5c3b4ed9d6f3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.932498] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 1017.932498] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e8edbc-1d44-5f83-aeeb-2a89cfa88198" [ 1017.932498] env[61867]: _type = "Task" [ 1017.932498] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.936791] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1017.936971] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1017.937985] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e719ebc3-2afd-4b34-a721-a30de1b2b41f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.943207] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e8edbc-1d44-5f83-aeeb-2a89cfa88198, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.946226] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1017.946226] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527f4b88-5a2a-a8c3-b7c2-425b4ffb5e9a" [ 1017.946226] env[61867]: _type = "Task" [ 1017.946226] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.953036] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527f4b88-5a2a-a8c3-b7c2-425b4ffb5e9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.064952] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277073, 'name': ReconfigVM_Task, 'duration_secs': 0.927907} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.065268] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 29d961df-5db5-46db-ace9-651913e0250a/29d961df-5db5-46db-ace9-651913e0250a.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1018.065874] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d9d0f3c-9549-407c-9133-062e4d63838d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.071892] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 1018.071892] env[61867]: value = "task-1277074" [ 1018.071892] env[61867]: _type = "Task" [ 1018.071892] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.080415] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277074, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.120192] env[61867]: DEBUG oslo_vmware.api [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.218697] env[61867]: DEBUG oslo_concurrency.lockutils [req-d127d5b5-d0b0-4534-8939-b16e627c9564 req-4eda00b4-f88c-46f3-b812-d58bbe8a2c2e service nova] Releasing lock "refresh_cache-986ada58-eb26-4289-bc0e-338c6a399645" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.388890] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6bc5b87a-dc4d-4b6d-b7a5-5c19214798b1 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "748e8a71-28b6-4c91-9cbb-856359348c96" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.095s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.443697] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e8edbc-1d44-5f83-aeeb-2a89cfa88198, 'name': SearchDatastore_Task, 'duration_secs': 0.00912} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.444012] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.444260] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.444471] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.454189] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527f4b88-5a2a-a8c3-b7c2-425b4ffb5e9a, 'name': SearchDatastore_Task, 'duration_secs': 0.009213} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.454843] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5a5bee2-a5e1-425c-96b1-1e82f0666bf4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.459361] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1018.459361] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523b2629-3f1e-adc7-3acd-c462d7ece527" [ 1018.459361] env[61867]: _type = "Task" [ 1018.459361] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.466009] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523b2629-3f1e-adc7-3acd-c462d7ece527, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.582389] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277074, 'name': Rename_Task, 'duration_secs': 0.312614} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.582671] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1018.582977] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a2a4bd49-4a7f-4c4b-a37c-0886b587cc1e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.589011] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 1018.589011] env[61867]: value = "task-1277075" [ 1018.589011] env[61867]: _type = "Task" [ 1018.589011] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.595994] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277075, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.619175] env[61867]: DEBUG oslo_vmware.api [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277060, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.970021] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523b2629-3f1e-adc7-3acd-c462d7ece527, 'name': SearchDatastore_Task, 'duration_secs': 0.008141} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.972273] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.972273] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 64947dbe-062c-4041-9ba9-d055279ded1e/64947dbe-062c-4041-9ba9-d055279ded1e.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1018.972273] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.972273] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.972273] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-602494e9-27ec-4961-b8c1-f56de6bbe69a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.973773] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-010dd753-72c7-47c0-889b-55365dedc27a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.975685] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1018.975932] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1018.980499] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1018.980499] env[61867]: value = "task-1277076" [ 1018.980499] env[61867]: _type = "Task" [ 1018.980499] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.984238] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.984431] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1018.985450] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07ed6bbf-290b-496d-a590-a0e61182b091 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.990269] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277076, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.992911] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 1018.992911] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52cb1f91-606d-f16a-e01f-04c9da1ba046" [ 1018.992911] env[61867]: _type = "Task" [ 1018.992911] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.999652] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52cb1f91-606d-f16a-e01f-04c9da1ba046, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.039330] env[61867]: DEBUG nova.compute.manager [req-bace6d8b-e104-46be-b76f-8b0aaa74e36d req-cf94c692-7066-495f-8e38-a4a3d5892b98 service nova] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Received event network-changed-eb146e5d-990b-477a-928e-ba9014b0fb40 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1019.039637] env[61867]: DEBUG nova.compute.manager [req-bace6d8b-e104-46be-b76f-8b0aaa74e36d req-cf94c692-7066-495f-8e38-a4a3d5892b98 service nova] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Refreshing instance network info cache due to event network-changed-eb146e5d-990b-477a-928e-ba9014b0fb40. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1019.039953] env[61867]: DEBUG oslo_concurrency.lockutils [req-bace6d8b-e104-46be-b76f-8b0aaa74e36d req-cf94c692-7066-495f-8e38-a4a3d5892b98 service nova] Acquiring lock "refresh_cache-748e8a71-28b6-4c91-9cbb-856359348c96" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.040223] env[61867]: DEBUG oslo_concurrency.lockutils [req-bace6d8b-e104-46be-b76f-8b0aaa74e36d req-cf94c692-7066-495f-8e38-a4a3d5892b98 service nova] Acquired lock "refresh_cache-748e8a71-28b6-4c91-9cbb-856359348c96" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.040495] env[61867]: DEBUG nova.network.neutron [req-bace6d8b-e104-46be-b76f-8b0aaa74e36d req-cf94c692-7066-495f-8e38-a4a3d5892b98 service nova] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Refreshing network info cache for port eb146e5d-990b-477a-928e-ba9014b0fb40 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1019.099073] env[61867]: DEBUG oslo_vmware.api [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277075, 'name': PowerOnVM_Task, 'duration_secs': 0.508503} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.099394] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1019.099642] env[61867]: INFO nova.compute.manager [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Took 8.42 seconds to spawn the instance on the hypervisor. [ 1019.099872] env[61867]: DEBUG nova.compute.manager [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1019.100712] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec291df8-d0cd-454f-83e0-dc38cfdd417e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.119872] env[61867]: DEBUG oslo_vmware.api [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277060, 'name': ReconfigVM_Task, 'duration_secs': 5.743876} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.120151] env[61867]: DEBUG oslo_concurrency.lockutils [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.120364] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Reconfigured VM to detach interface {{(pid=61867) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1019.492281] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277076, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451871} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.492560] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 64947dbe-062c-4041-9ba9-d055279ded1e/64947dbe-062c-4041-9ba9-d055279ded1e.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1019.492796] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1019.493114] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b72672ab-c10d-457e-898d-597e0cb1f62d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.504191] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52cb1f91-606d-f16a-e01f-04c9da1ba046, 'name': SearchDatastore_Task, 'duration_secs': 0.007991} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.508087] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1019.508087] env[61867]: value = "task-1277077" [ 1019.508087] env[61867]: _type = "Task" [ 1019.508087] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.508087] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1b4046a-764b-4c96-ba87-fb96721e82cb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.514707] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 1019.514707] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b62366-de35-fe49-1b4c-742477e7863b" [ 1019.514707] env[61867]: _type = "Task" [ 1019.514707] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.517925] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277077, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.525849] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b62366-de35-fe49-1b4c-742477e7863b, 'name': SearchDatastore_Task, 'duration_secs': 0.008718} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.526141] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.526418] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 986ada58-eb26-4289-bc0e-338c6a399645/986ada58-eb26-4289-bc0e-338c6a399645.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1019.526674] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e6fc7e60-f4ef-4223-bcea-f606956a987e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.533210] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 1019.533210] env[61867]: value = "task-1277078" [ 1019.533210] env[61867]: _type = "Task" [ 1019.533210] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.541434] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277078, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.621619] env[61867]: INFO nova.compute.manager [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Took 18.00 seconds to build instance. [ 1019.789995] env[61867]: DEBUG nova.network.neutron [req-bace6d8b-e104-46be-b76f-8b0aaa74e36d req-cf94c692-7066-495f-8e38-a4a3d5892b98 service nova] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Updated VIF entry in instance network info cache for port eb146e5d-990b-477a-928e-ba9014b0fb40. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1019.790416] env[61867]: DEBUG nova.network.neutron [req-bace6d8b-e104-46be-b76f-8b0aaa74e36d req-cf94c692-7066-495f-8e38-a4a3d5892b98 service nova] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Updating instance_info_cache with network_info: [{"id": "eb146e5d-990b-477a-928e-ba9014b0fb40", "address": "fa:16:3e:34:92:d2", "network": {"id": "fc403caf-421f-4783-9807-5e61c4379a6d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-846804042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24180ce8aca142fb897e29d4853c20c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb146e5d-99", "ovs_interfaceid": "eb146e5d-990b-477a-928e-ba9014b0fb40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.018511] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277077, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06131} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.018930] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1020.019770] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7725c72a-b0d5-4e05-bb5f-15644b5cd4e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.039292] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 64947dbe-062c-4041-9ba9-d055279ded1e/64947dbe-062c-4041-9ba9-d055279ded1e.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1020.039915] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e472094-93dc-4f06-b2a7-a071994c0771 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.062961] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277078, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471452} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.064222] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 986ada58-eb26-4289-bc0e-338c6a399645/986ada58-eb26-4289-bc0e-338c6a399645.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1020.064526] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1020.065612] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1020.065612] env[61867]: value = "task-1277079" [ 1020.065612] env[61867]: _type = "Task" [ 1020.065612] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.065612] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-32dbee94-4fb2-4a6e-a328-9c5e7b3b512b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.074707] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277079, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.075778] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 1020.075778] env[61867]: value = "task-1277080" [ 1020.075778] env[61867]: _type = "Task" [ 1020.075778] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.084931] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277080, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.124326] env[61867]: DEBUG oslo_concurrency.lockutils [None req-766d33dc-e939-41bb-803e-516c9922b0c0 tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "29d961df-5db5-46db-ace9-651913e0250a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.513s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.294064] env[61867]: DEBUG oslo_concurrency.lockutils [req-bace6d8b-e104-46be-b76f-8b0aaa74e36d req-cf94c692-7066-495f-8e38-a4a3d5892b98 service nova] Releasing lock "refresh_cache-748e8a71-28b6-4c91-9cbb-856359348c96" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.461620] env[61867]: DEBUG oslo_concurrency.lockutils [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.461735] env[61867]: DEBUG oslo_concurrency.lockutils [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquired lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.461857] env[61867]: DEBUG nova.network.neutron [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1020.577544] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277079, 'name': ReconfigVM_Task, 'duration_secs': 0.286369} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.580759] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 64947dbe-062c-4041-9ba9-d055279ded1e/64947dbe-062c-4041-9ba9-d055279ded1e.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1020.581400] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5e435dfb-3716-4e3a-9a24-9aba38ed29fe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.588083] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277080, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069591} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.589387] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1020.589743] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1020.589743] env[61867]: value = "task-1277081" [ 1020.589743] env[61867]: _type = "Task" [ 1020.589743] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.590414] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffcd5bfb-c4d1-4fba-a439-ee61d9fc4992 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.609605] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277081, 'name': Rename_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.618680] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 986ada58-eb26-4289-bc0e-338c6a399645/986ada58-eb26-4289-bc0e-338c6a399645.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1020.618985] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a468fbde-44d1-4895-8c12-f471947db027 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.637727] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 1020.637727] env[61867]: value = "task-1277082" [ 1020.637727] env[61867]: _type = "Task" [ 1020.637727] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.645677] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277082, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.932691] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "0ae83268-ef50-42e6-954d-5e67c4afc20e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.933051] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "0ae83268-ef50-42e6-954d-5e67c4afc20e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.933285] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "0ae83268-ef50-42e6-954d-5e67c4afc20e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.933481] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "0ae83268-ef50-42e6-954d-5e67c4afc20e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.933660] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "0ae83268-ef50-42e6-954d-5e67c4afc20e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.935925] env[61867]: INFO nova.compute.manager [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Terminating instance [ 1020.937697] env[61867]: DEBUG nova.compute.manager [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1020.937895] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1020.938989] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed8b603-ff06-453b-a08f-c296d95cfd00 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.946709] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1020.946957] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c3420f1-7d9c-4073-8d6a-bdbde2a86391 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.953321] env[61867]: DEBUG oslo_vmware.api [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 1020.953321] env[61867]: value = "task-1277083" [ 1020.953321] env[61867]: _type = "Task" [ 1020.953321] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.960980] env[61867]: DEBUG oslo_vmware.api [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277083, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.101894] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277081, 'name': Rename_Task, 'duration_secs': 0.137481} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.104425] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1021.104684] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d9f9121-c879-4e78-94ed-232c2cead75c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.110698] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1021.110698] env[61867]: value = "task-1277084" [ 1021.110698] env[61867]: _type = "Task" [ 1021.110698] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.117790] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277084, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.147650] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277082, 'name': ReconfigVM_Task, 'duration_secs': 0.260267} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.147650] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 986ada58-eb26-4289-bc0e-338c6a399645/986ada58-eb26-4289-bc0e-338c6a399645.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1021.148289] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-28cafdde-9fe0-45c0-8902-329792d9c317 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.155698] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 1021.155698] env[61867]: value = "task-1277085" [ 1021.155698] env[61867]: _type = "Task" [ 1021.155698] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.163543] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277085, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.176623] env[61867]: INFO nova.network.neutron [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Port aa5b1005-fdcb-4208-b0d8-3b33573d05b6 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1021.176974] env[61867]: DEBUG nova.network.neutron [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Updating instance_info_cache with network_info: [{"id": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "address": "fa:16:3e:a0:23:06", "network": {"id": "7a62cd67-4e7d-4952-9a24-053e35c1c3cb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1601454180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86ced20eaf4740e298dc6f8ca5550c09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11da2092-76f7-447e-babb-8fc14ad39a71", "external-id": "nsx-vlan-transportzone-585", "segmentation_id": 585, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3537cac-b5", "ovs_interfaceid": "a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.308907] env[61867]: DEBUG oslo_concurrency.lockutils [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "29d961df-5db5-46db-ace9-651913e0250a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.309226] env[61867]: DEBUG oslo_concurrency.lockutils [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "29d961df-5db5-46db-ace9-651913e0250a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.309454] env[61867]: DEBUG oslo_concurrency.lockutils [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "29d961df-5db5-46db-ace9-651913e0250a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.309647] env[61867]: DEBUG oslo_concurrency.lockutils [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "29d961df-5db5-46db-ace9-651913e0250a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.309829] env[61867]: DEBUG oslo_concurrency.lockutils [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "29d961df-5db5-46db-ace9-651913e0250a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.312514] env[61867]: INFO nova.compute.manager [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Terminating instance [ 1021.314324] env[61867]: DEBUG nova.compute.manager [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1021.314515] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1021.315345] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a09b5f-d05e-41ad-bdc2-a3d6598ac85b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.324936] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1021.325145] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-43ec8d60-4002-4194-8c7d-12ab9327f9c5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.337640] env[61867]: DEBUG oslo_vmware.api [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 1021.337640] env[61867]: value = "task-1277086" [ 1021.337640] env[61867]: _type = "Task" [ 1021.337640] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.345604] env[61867]: DEBUG oslo_vmware.api [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277086, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.463812] env[61867]: DEBUG oslo_vmware.api [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277083, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.619911] env[61867]: DEBUG oslo_vmware.api [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277084, 'name': PowerOnVM_Task, 'duration_secs': 0.449598} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.620203] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1021.620410] env[61867]: INFO nova.compute.manager [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Took 4.82 seconds to spawn the instance on the hypervisor. [ 1021.620596] env[61867]: DEBUG nova.compute.manager [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1021.621387] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd96c6b3-526b-4260-8047-730a7d6da4fe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.666131] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277085, 'name': Rename_Task, 'duration_secs': 0.174539} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.666510] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1021.666834] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f2348e4a-fcb9-4f00-b157-f87a7ae79c34 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.673023] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 1021.673023] env[61867]: value = "task-1277087" [ 1021.673023] env[61867]: _type = "Task" [ 1021.673023] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.680817] env[61867]: DEBUG oslo_concurrency.lockutils [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Releasing lock "refresh_cache-0ae83268-ef50-42e6-954d-5e67c4afc20e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.682843] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277087, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.847518] env[61867]: DEBUG oslo_vmware.api [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277086, 'name': PowerOffVM_Task, 'duration_secs': 0.340013} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.847799] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1021.847978] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1021.848257] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-283f0f4f-7172-4132-872d-adbe8a90b82a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.910515] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1021.910763] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1021.911127] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleting the datastore file [datastore2] 29d961df-5db5-46db-ace9-651913e0250a {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1021.911306] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-be4e0783-7b38-489d-bd13-371cda7c7fa3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.917590] env[61867]: DEBUG oslo_vmware.api [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for the task: (returnval){ [ 1021.917590] env[61867]: value = "task-1277089" [ 1021.917590] env[61867]: _type = "Task" [ 1021.917590] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.926249] env[61867]: DEBUG oslo_vmware.api [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277089, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.963971] env[61867]: DEBUG oslo_vmware.api [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277083, 'name': PowerOffVM_Task, 'duration_secs': 0.713991} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.964262] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1021.964436] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1021.964683] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b365b6ca-c9f1-43f2-bb4f-a5e65766f0fb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.025053] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1022.025304] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1022.025491] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Deleting the datastore file [datastore2] 0ae83268-ef50-42e6-954d-5e67c4afc20e {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1022.025767] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1bbaa586-0265-40e1-acaf-bf6088943408 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.032308] env[61867]: DEBUG oslo_vmware.api [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 1022.032308] env[61867]: value = "task-1277091" [ 1022.032308] env[61867]: _type = "Task" [ 1022.032308] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.039865] env[61867]: DEBUG oslo_vmware.api [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277091, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.138429] env[61867]: INFO nova.compute.manager [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Took 14.15 seconds to build instance. [ 1022.185375] env[61867]: DEBUG oslo_concurrency.lockutils [None req-87eaa0bd-922c-403d-a3ea-6076fac3149a tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "interface-0ae83268-ef50-42e6-954d-5e67c4afc20e-aa5b1005-fdcb-4208-b0d8-3b33573d05b6" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.658s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.186694] env[61867]: DEBUG oslo_vmware.api [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277087, 'name': PowerOnVM_Task, 'duration_secs': 0.495867} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.187187] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1022.187447] env[61867]: INFO nova.compute.manager [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Took 7.59 seconds to spawn the instance on the hypervisor. [ 1022.187681] env[61867]: DEBUG nova.compute.manager [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.188516] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-271e7cc4-46eb-4629-ba97-129c996dd3e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.427478] env[61867]: DEBUG oslo_vmware.api [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Task: {'id': task-1277089, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140893} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.427683] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1022.427872] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1022.428067] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1022.428250] env[61867]: INFO nova.compute.manager [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1022.428491] env[61867]: DEBUG oslo.service.loopingcall [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.428682] env[61867]: DEBUG nova.compute.manager [-] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1022.428776] env[61867]: DEBUG nova.network.neutron [-] [instance: 29d961df-5db5-46db-ace9-651913e0250a] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1022.446961] env[61867]: INFO nova.compute.manager [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Rebuilding instance [ 1022.483551] env[61867]: DEBUG nova.compute.manager [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.484411] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e136689c-2f2b-47e5-bb3b-390e2de78ab2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.541788] env[61867]: DEBUG oslo_vmware.api [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277091, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.202993} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.542112] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1022.542312] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1022.542489] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1022.542660] env[61867]: INFO nova.compute.manager [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Took 1.60 seconds to destroy the instance on the hypervisor. [ 1022.542958] env[61867]: DEBUG oslo.service.loopingcall [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.543160] env[61867]: DEBUG nova.compute.manager [-] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1022.543254] env[61867]: DEBUG nova.network.neutron [-] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1022.640873] env[61867]: DEBUG oslo_concurrency.lockutils [None req-2b1d126f-6021-4069-9ea9-c31e21ccf4fe tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Lock "64947dbe-062c-4041-9ba9-d055279ded1e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.668s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.706307] env[61867]: INFO nova.compute.manager [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Took 16.18 seconds to build instance. [ 1022.760805] env[61867]: DEBUG nova.compute.manager [req-1c0d472c-8f3f-40eb-bb72-1f99adedce80 req-4f15e994-087f-4e88-a55f-22d361681fe6 service nova] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Received event network-vif-deleted-380ad3aa-f8aa-41b7-94bb-71a4528022f2 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1022.761037] env[61867]: INFO nova.compute.manager [req-1c0d472c-8f3f-40eb-bb72-1f99adedce80 req-4f15e994-087f-4e88-a55f-22d361681fe6 service nova] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Neutron deleted interface 380ad3aa-f8aa-41b7-94bb-71a4528022f2; detaching it from the instance and deleting it from the info cache [ 1022.761219] env[61867]: DEBUG nova.network.neutron [req-1c0d472c-8f3f-40eb-bb72-1f99adedce80 req-4f15e994-087f-4e88-a55f-22d361681fe6 service nova] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.996580] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1022.996966] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dfcbfe55-c825-44ec-a20a-4329c540a221 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.003984] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1023.003984] env[61867]: value = "task-1277092" [ 1023.003984] env[61867]: _type = "Task" [ 1023.003984] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.012513] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277092, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.208544] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b538c16e-c01d-47e0-a07f-3e48b9c60998 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "986ada58-eb26-4289-bc0e-338c6a399645" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.686s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.242166] env[61867]: DEBUG nova.network.neutron [-] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.265423] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0f57f895-c2d4-4d73-8a8c-135023073c94 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.275328] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ca26b3-026c-48b3-a6e0-c1de2356a226 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.302959] env[61867]: DEBUG nova.compute.manager [req-1c0d472c-8f3f-40eb-bb72-1f99adedce80 req-4f15e994-087f-4e88-a55f-22d361681fe6 service nova] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Detach interface failed, port_id=380ad3aa-f8aa-41b7-94bb-71a4528022f2, reason: Instance 29d961df-5db5-46db-ace9-651913e0250a could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1023.513701] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277092, 'name': PowerOffVM_Task, 'duration_secs': 0.23346} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.513970] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1023.514210] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1023.514952] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69b307f-6d60-43eb-a6e7-e9498fda0261 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.521100] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1023.521299] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8d31f82b-9c2b-47f9-9369-6ec25e22a052 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.543472] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1023.543699] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1023.543928] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Deleting the datastore file [datastore2] 64947dbe-062c-4041-9ba9-d055279ded1e {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1023.544181] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fb746391-5bda-440f-8a88-cede21f4719c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.550819] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1023.550819] env[61867]: value = "task-1277094" [ 1023.550819] env[61867]: _type = "Task" [ 1023.550819] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.557923] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277094, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.739674] env[61867]: DEBUG nova.network.neutron [-] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.744967] env[61867]: INFO nova.compute.manager [-] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Took 1.32 seconds to deallocate network for instance. [ 1024.060353] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277094, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.085507} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.060567] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1024.060772] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1024.060960] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1024.216960] env[61867]: DEBUG nova.compute.manager [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Stashing vm_state: active {{(pid=61867) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1024.243080] env[61867]: INFO nova.compute.manager [-] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Took 1.70 seconds to deallocate network for instance. [ 1024.251382] env[61867]: DEBUG oslo_concurrency.lockutils [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.251382] env[61867]: DEBUG oslo_concurrency.lockutils [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.251515] env[61867]: DEBUG nova.objects.instance [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lazy-loading 'resources' on Instance uuid 29d961df-5db5-46db-ace9-651913e0250a {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.737979] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.748578] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.792287] env[61867]: DEBUG nova.compute.manager [req-6822e197-3413-4435-ba96-42027090fbd2 req-5fda8909-17b2-4d4e-b6bb-e37947bde7d7 service nova] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Received event network-vif-deleted-a3537cac-b5b2-4caf-87b6-ccd7ca1f6dab {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1024.858646] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c0a512-4b5c-463e-8bb4-500a4aa66bf7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.866720] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1f6e20-7bb8-478e-9b11-5eead39cfbd9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.911283] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-536f6781-d329-4071-8186-d6d36bba9854 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.921896] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ceed014-6bc4-41b1-9a80-033433a962b1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.943100] env[61867]: DEBUG nova.compute.provider_tree [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.096437] env[61867]: DEBUG nova.virt.hardware [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1025.096692] env[61867]: DEBUG nova.virt.hardware [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1025.096855] env[61867]: DEBUG nova.virt.hardware [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1025.097052] env[61867]: DEBUG nova.virt.hardware [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1025.097215] env[61867]: DEBUG nova.virt.hardware [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1025.097366] env[61867]: DEBUG nova.virt.hardware [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1025.097574] env[61867]: DEBUG nova.virt.hardware [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1025.097739] env[61867]: DEBUG nova.virt.hardware [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1025.097907] env[61867]: DEBUG nova.virt.hardware [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1025.098084] env[61867]: DEBUG nova.virt.hardware [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1025.098266] env[61867]: DEBUG nova.virt.hardware [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1025.099115] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd18052-bd78-4637-8c91-d78f5efbe1fb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.106870] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a291fb2-5293-48c5-b41d-d9c4eb457af9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.120011] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Instance VIF info [] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1025.125514] env[61867]: DEBUG oslo.service.loopingcall [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1025.125720] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1025.125929] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-53be9013-a219-490b-847b-1fb08e1259a5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.141815] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1025.141815] env[61867]: value = "task-1277095" [ 1025.141815] env[61867]: _type = "Task" [ 1025.141815] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.149261] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277095, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.447165] env[61867]: DEBUG nova.scheduler.client.report [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1025.652612] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277095, 'name': CreateVM_Task, 'duration_secs': 0.249305} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.652765] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1025.653228] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.653413] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.653745] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1025.654126] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c405fcb2-3641-4747-8bdc-d1ff27b0f2d5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.658706] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1025.658706] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b139c3-e2e6-e637-f89e-2eb4cf837a00" [ 1025.658706] env[61867]: _type = "Task" [ 1025.658706] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.666259] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b139c3-e2e6-e637-f89e-2eb4cf837a00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.952076] env[61867]: DEBUG oslo_concurrency.lockutils [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.701s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.954618] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.217s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.970939] env[61867]: INFO nova.scheduler.client.report [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Deleted allocations for instance 29d961df-5db5-46db-ace9-651913e0250a [ 1026.168889] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b139c3-e2e6-e637-f89e-2eb4cf837a00, 'name': SearchDatastore_Task, 'duration_secs': 0.045909} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.169224] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.169470] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1026.169705] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.169855] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.170055] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1026.170345] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e86418fa-67b4-4c05-b967-66dbf0805426 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.182100] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1026.182308] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1026.183131] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-913ab9f5-c9bb-450d-a373-4d12dc0f5e63 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.188837] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1026.188837] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52aceb9a-8b61-838b-6d8c-1f93ed9e62eb" [ 1026.188837] env[61867]: _type = "Task" [ 1026.188837] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.196566] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52aceb9a-8b61-838b-6d8c-1f93ed9e62eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.460361] env[61867]: INFO nova.compute.claims [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1026.477766] env[61867]: DEBUG oslo_concurrency.lockutils [None req-05e3dea6-45e5-4ce0-bfc3-5593f27c01ad tempest-ServerDiskConfigTestJSON-983631310 tempest-ServerDiskConfigTestJSON-983631310-project-member] Lock "29d961df-5db5-46db-ace9-651913e0250a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.168s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.710346] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52aceb9a-8b61-838b-6d8c-1f93ed9e62eb, 'name': SearchDatastore_Task, 'duration_secs': 0.025533} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.711677] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c48e6896-e8fd-4a1f-af6d-044bea7eeaaa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.716764] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1026.716764] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523be5cf-f5b4-bac6-5cdb-dcfe264b4007" [ 1026.716764] env[61867]: _type = "Task" [ 1026.716764] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.725766] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523be5cf-f5b4-bac6-5cdb-dcfe264b4007, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.968087] env[61867]: INFO nova.compute.resource_tracker [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Updating resource usage from migration 2ffb5793-9433-4a7a-b35c-ee8968f54a67 [ 1027.079901] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-440e7d48-02da-4bea-b602-26d0ff6524e3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.092288] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d334bcf3-3291-4b34-893d-c80662995eef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.147266] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32981edd-c954-46c7-91bf-e9b545703b06 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.155540] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0963b0a3-8568-4d89-b3f5-8f8466540ee3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.171421] env[61867]: DEBUG nova.compute.provider_tree [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.227930] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523be5cf-f5b4-bac6-5cdb-dcfe264b4007, 'name': SearchDatastore_Task, 'duration_secs': 0.011029} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.228205] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.228555] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 64947dbe-062c-4041-9ba9-d055279ded1e/64947dbe-062c-4041-9ba9-d055279ded1e.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1027.228767] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e178453d-e568-4207-8099-89c558393743 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.235757] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1027.235757] env[61867]: value = "task-1277096" [ 1027.235757] env[61867]: _type = "Task" [ 1027.235757] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.245737] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277096, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.676207] env[61867]: DEBUG nova.scheduler.client.report [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1027.747070] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277096, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.183158] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.228s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.183402] env[61867]: INFO nova.compute.manager [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Migrating [ 1028.190037] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.442s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.194017] env[61867]: DEBUG nova.objects.instance [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lazy-loading 'resources' on Instance uuid 0ae83268-ef50-42e6-954d-5e67c4afc20e {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.246235] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277096, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.667267} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.246523] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 64947dbe-062c-4041-9ba9-d055279ded1e/64947dbe-062c-4041-9ba9-d055279ded1e.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1028.246752] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1028.247271] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2c41db95-3671-4438-aef8-42cbe5ad21fc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.254292] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1028.254292] env[61867]: value = "task-1277097" [ 1028.254292] env[61867]: _type = "Task" [ 1028.254292] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.262349] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277097, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.701399] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "refresh_cache-986ada58-eb26-4289-bc0e-338c6a399645" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.701597] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "refresh_cache-986ada58-eb26-4289-bc0e-338c6a399645" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.701752] env[61867]: DEBUG nova.network.neutron [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1028.765263] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277097, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065636} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.765595] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1028.766409] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbbafb8b-b383-432f-bf1f-2f790b8ac62b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.788142] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 64947dbe-062c-4041-9ba9-d055279ded1e/64947dbe-062c-4041-9ba9-d055279ded1e.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1028.790820] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f38c1906-2682-4f52-b626-2ecdd87b4817 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.812134] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1028.812134] env[61867]: value = "task-1277098" [ 1028.812134] env[61867]: _type = "Task" [ 1028.812134] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.817223] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e65d0a6-511f-4261-adcf-86f550979fd6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.822799] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277098, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.827246] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-162eb433-3222-4d62-af03-5f43c02deb8d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.857799] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630d18c1-8fc4-445a-ab28-cd68507e4d5e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.864709] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b3bef57-e23a-4509-b146-f36646fc5ea1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.877709] env[61867]: DEBUG nova.compute.provider_tree [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1029.322705] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277098, 'name': ReconfigVM_Task, 'duration_secs': 0.272879} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.323031] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 64947dbe-062c-4041-9ba9-d055279ded1e/64947dbe-062c-4041-9ba9-d055279ded1e.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1029.323658] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5b20e0fe-85eb-4d8b-a1d0-7806290367e2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.329404] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1029.329404] env[61867]: value = "task-1277099" [ 1029.329404] env[61867]: _type = "Task" [ 1029.329404] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.336792] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277099, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.380703] env[61867]: DEBUG nova.scheduler.client.report [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.420471] env[61867]: DEBUG nova.network.neutron [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Updating instance_info_cache with network_info: [{"id": "072ec5db-d2ce-4e94-809d-7ff5fd034562", "address": "fa:16:3e:45:de:03", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap072ec5db-d2", "ovs_interfaceid": "072ec5db-d2ce-4e94-809d-7ff5fd034562", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.839972] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277099, 'name': Rename_Task, 'duration_secs': 0.130825} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.840634] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1029.840925] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a4bf8f30-b25d-44bd-a5e6-eaf85693b69f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.847124] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1029.847124] env[61867]: value = "task-1277100" [ 1029.847124] env[61867]: _type = "Task" [ 1029.847124] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.854588] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277100, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.885643] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.696s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.918625] env[61867]: INFO nova.scheduler.client.report [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Deleted allocations for instance 0ae83268-ef50-42e6-954d-5e67c4afc20e [ 1029.924273] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "refresh_cache-986ada58-eb26-4289-bc0e-338c6a399645" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.357451] env[61867]: DEBUG oslo_vmware.api [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277100, 'name': PowerOnVM_Task, 'duration_secs': 0.47256} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.357747] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1030.357959] env[61867]: DEBUG nova.compute.manager [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1030.358743] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad8ec755-bac8-4239-bffc-8cda33affe25 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.430538] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6322d0e1-73c1-462c-9e19-4ce76d033171 tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "0ae83268-ef50-42e6-954d-5e67c4afc20e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.497s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.829151] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "4d7bfb31-d565-49e4-8c51-0122acebba2d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.829461] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "4d7bfb31-d565-49e4-8c51-0122acebba2d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.829679] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "4d7bfb31-d565-49e4-8c51-0122acebba2d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.829873] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "4d7bfb31-d565-49e4-8c51-0122acebba2d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.830065] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "4d7bfb31-d565-49e4-8c51-0122acebba2d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.832294] env[61867]: INFO nova.compute.manager [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Terminating instance [ 1030.834553] env[61867]: DEBUG nova.compute.manager [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1030.834742] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1030.835638] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d0f537f-15c0-4516-8513-543de3c0910a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.843043] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1030.843309] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e1d24215-96a7-4671-a24f-f1e4c592a00f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.849102] env[61867]: DEBUG oslo_vmware.api [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 1030.849102] env[61867]: value = "task-1277101" [ 1030.849102] env[61867]: _type = "Task" [ 1030.849102] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.857083] env[61867]: DEBUG oslo_vmware.api [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277101, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.878028] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.878496] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.878606] env[61867]: DEBUG nova.objects.instance [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61867) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1031.111623] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquiring lock "64947dbe-062c-4041-9ba9-d055279ded1e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.112150] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Lock "64947dbe-062c-4041-9ba9-d055279ded1e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.112481] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquiring lock "64947dbe-062c-4041-9ba9-d055279ded1e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.112987] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Lock "64947dbe-062c-4041-9ba9-d055279ded1e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.113306] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Lock "64947dbe-062c-4041-9ba9-d055279ded1e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.115986] env[61867]: INFO nova.compute.manager [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Terminating instance [ 1031.117763] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquiring lock "refresh_cache-64947dbe-062c-4041-9ba9-d055279ded1e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.117917] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquired lock "refresh_cache-64947dbe-062c-4041-9ba9-d055279ded1e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.118996] env[61867]: DEBUG nova.network.neutron [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1031.360289] env[61867]: DEBUG oslo_vmware.api [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277101, 'name': PowerOffVM_Task, 'duration_secs': 0.208229} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.360289] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1031.360619] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1031.361119] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d77ceb49-125a-4b6f-a336-49cd4cf690ab {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.435130] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1031.435401] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1031.435592] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Deleting the datastore file [datastore1] 4d7bfb31-d565-49e4-8c51-0122acebba2d {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1031.435874] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f319d5aa-9b87-4746-a002-41bd2d849fd0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.442176] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f23d42ef-ab96-45ab-899d-8b7c3f1d0013 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.448092] env[61867]: DEBUG oslo_vmware.api [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for the task: (returnval){ [ 1031.448092] env[61867]: value = "task-1277103" [ 1031.448092] env[61867]: _type = "Task" [ 1031.448092] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.468339] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Updating instance '986ada58-eb26-4289-bc0e-338c6a399645' progress to 0 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1031.477714] env[61867]: DEBUG oslo_vmware.api [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277103, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.640700] env[61867]: DEBUG nova.network.neutron [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1031.700772] env[61867]: DEBUG nova.network.neutron [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.888287] env[61867]: DEBUG oslo_concurrency.lockutils [None req-3f947abc-f573-4254-aa3f-03c9bee99f20 tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.931284] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "a9724422-abd1-4771-acb2-2c0ac45f5c16" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.931576] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "a9724422-abd1-4771-acb2-2c0ac45f5c16" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.957650] env[61867]: DEBUG oslo_vmware.api [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Task: {'id': task-1277103, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156116} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.958041] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1031.958180] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1031.958435] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1031.958534] env[61867]: INFO nova.compute.manager [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1031.958783] env[61867]: DEBUG oslo.service.loopingcall [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1031.959025] env[61867]: DEBUG nova.compute.manager [-] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1031.959084] env[61867]: DEBUG nova.network.neutron [-] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1031.974389] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1031.974687] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c84e3398-3dbc-4aec-b047-46c3fa218009 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.981734] env[61867]: DEBUG oslo_vmware.api [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 1031.981734] env[61867]: value = "task-1277104" [ 1031.981734] env[61867]: _type = "Task" [ 1031.981734] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.990851] env[61867]: DEBUG oslo_vmware.api [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277104, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.204129] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Releasing lock "refresh_cache-64947dbe-062c-4041-9ba9-d055279ded1e" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.204608] env[61867]: DEBUG nova.compute.manager [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1032.204808] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1032.206058] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d1f35f-75de-439c-828c-1762cdcfc31e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.215147] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1032.215346] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-75d20b90-536e-4cfa-9947-f967042cf396 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.221231] env[61867]: DEBUG oslo_vmware.api [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1032.221231] env[61867]: value = "task-1277105" [ 1032.221231] env[61867]: _type = "Task" [ 1032.221231] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.229882] env[61867]: DEBUG oslo_vmware.api [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277105, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.278556] env[61867]: DEBUG nova.compute.manager [req-ff75e3d5-5236-4121-baa7-17678e49ce7d req-1b8bd496-e221-4b39-a692-063527583f4e service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Received event network-vif-deleted-c05702c4-5138-450d-82b2-790b86d8a59b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1032.278775] env[61867]: INFO nova.compute.manager [req-ff75e3d5-5236-4121-baa7-17678e49ce7d req-1b8bd496-e221-4b39-a692-063527583f4e service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Neutron deleted interface c05702c4-5138-450d-82b2-790b86d8a59b; detaching it from the instance and deleting it from the info cache [ 1032.278907] env[61867]: DEBUG nova.network.neutron [req-ff75e3d5-5236-4121-baa7-17678e49ce7d req-1b8bd496-e221-4b39-a692-063527583f4e service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.434022] env[61867]: DEBUG nova.compute.manager [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1032.493348] env[61867]: DEBUG oslo_vmware.api [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277104, 'name': PowerOffVM_Task, 'duration_secs': 0.202389} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.493607] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1032.493797] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Updating instance '986ada58-eb26-4289-bc0e-338c6a399645' progress to 17 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1032.732387] env[61867]: DEBUG oslo_vmware.api [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277105, 'name': PowerOffVM_Task, 'duration_secs': 0.114235} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.732655] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1032.732834] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1032.733180] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7b09569e-d6ae-49a1-96bc-53753f68133f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.754345] env[61867]: DEBUG nova.network.neutron [-] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.758984] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1032.758984] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1032.758984] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Deleting the datastore file [datastore2] 64947dbe-062c-4041-9ba9-d055279ded1e {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1032.759398] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e5ca0539-81d2-4980-b426-e082758d8aa6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.766892] env[61867]: DEBUG oslo_vmware.api [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for the task: (returnval){ [ 1032.766892] env[61867]: value = "task-1277107" [ 1032.766892] env[61867]: _type = "Task" [ 1032.766892] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.777316] env[61867]: DEBUG oslo_vmware.api [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277107, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.782025] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0b607d6b-08e9-4af5-b5a7-ca3acf28f92a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.790165] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b88015-2825-4187-ba07-9a5b5509103e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.813979] env[61867]: DEBUG nova.compute.manager [req-ff75e3d5-5236-4121-baa7-17678e49ce7d req-1b8bd496-e221-4b39-a692-063527583f4e service nova] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Detach interface failed, port_id=c05702c4-5138-450d-82b2-790b86d8a59b, reason: Instance 4d7bfb31-d565-49e4-8c51-0122acebba2d could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1032.957101] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.957382] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.958864] env[61867]: INFO nova.compute.claims [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1032.999751] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1032.999984] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1033.000163] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1033.000351] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1033.000501] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1033.000651] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1033.001145] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1033.001145] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1033.001277] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1033.001513] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1033.001710] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1033.007080] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c074f57-04bf-487e-b3c0-71882e9e9830 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.023010] env[61867]: DEBUG oslo_vmware.api [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 1033.023010] env[61867]: value = "task-1277108" [ 1033.023010] env[61867]: _type = "Task" [ 1033.023010] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.034080] env[61867]: DEBUG oslo_vmware.api [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277108, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.256945] env[61867]: INFO nova.compute.manager [-] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Took 1.30 seconds to deallocate network for instance. [ 1033.276857] env[61867]: DEBUG oslo_vmware.api [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Task: {'id': task-1277107, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.086102} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.277123] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1033.277316] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1033.277497] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1033.277672] env[61867]: INFO nova.compute.manager [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Took 1.07 seconds to destroy the instance on the hypervisor. [ 1033.277908] env[61867]: DEBUG oslo.service.loopingcall [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1033.278121] env[61867]: DEBUG nova.compute.manager [-] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1033.278207] env[61867]: DEBUG nova.network.neutron [-] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1033.294604] env[61867]: DEBUG nova.network.neutron [-] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1033.532426] env[61867]: DEBUG oslo_vmware.api [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277108, 'name': ReconfigVM_Task, 'duration_secs': 0.187551} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.532849] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Updating instance '986ada58-eb26-4289-bc0e-338c6a399645' progress to 33 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1033.763830] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.796568] env[61867]: DEBUG nova.network.neutron [-] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.038617] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1034.038883] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1034.039030] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1034.039225] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1034.039376] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1034.039530] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1034.039741] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1034.039903] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1034.040095] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1034.040441] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1034.040665] env[61867]: DEBUG nova.virt.hardware [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1034.046428] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Reconfiguring VM instance instance-00000067 to detach disk 2000 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1034.046890] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d11d688-6109-4c42-a96a-ab2dfaf9a4b7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.066848] env[61867]: DEBUG oslo_vmware.api [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 1034.066848] env[61867]: value = "task-1277109" [ 1034.066848] env[61867]: _type = "Task" [ 1034.066848] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.071039] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec122bc9-44df-4241-9303-06f65deeb96c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.077895] env[61867]: DEBUG oslo_vmware.api [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277109, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.080288] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c026414c-fa8d-4ccb-a344-fea5551336df {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.109956] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9831949-2440-4979-a095-54afbd047195 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.117211] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f3794b-fb05-4ded-9832-0c1e23298faf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.131677] env[61867]: DEBUG nova.compute.provider_tree [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.299833] env[61867]: INFO nova.compute.manager [-] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Took 1.02 seconds to deallocate network for instance. [ 1034.576043] env[61867]: DEBUG oslo_vmware.api [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277109, 'name': ReconfigVM_Task, 'duration_secs': 0.152917} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.576350] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Reconfigured VM instance instance-00000067 to detach disk 2000 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1034.577115] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3319f715-e0cb-4093-80cb-a30ef8c838be {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.598154] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 986ada58-eb26-4289-bc0e-338c6a399645/986ada58-eb26-4289-bc0e-338c6a399645.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1034.598745] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a876476-59f4-4278-9677-a3874101a6d1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.615593] env[61867]: DEBUG oslo_vmware.api [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 1034.615593] env[61867]: value = "task-1277110" [ 1034.615593] env[61867]: _type = "Task" [ 1034.615593] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.622902] env[61867]: DEBUG oslo_vmware.api [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277110, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.634869] env[61867]: DEBUG nova.scheduler.client.report [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1034.807103] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.125162] env[61867]: DEBUG oslo_vmware.api [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277110, 'name': ReconfigVM_Task, 'duration_secs': 0.252109} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.125488] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 986ada58-eb26-4289-bc0e-338c6a399645/986ada58-eb26-4289-bc0e-338c6a399645.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1035.125735] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Updating instance '986ada58-eb26-4289-bc0e-338c6a399645' progress to 50 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1035.139297] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.182s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.139793] env[61867]: DEBUG nova.compute.manager [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1035.142225] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.379s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.142544] env[61867]: DEBUG nova.objects.instance [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lazy-loading 'resources' on Instance uuid 4d7bfb31-d565-49e4-8c51-0122acebba2d {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.207321] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.207510] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.632818] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a923f1-0ac7-4b50-b63f-eb9016cc73d6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.652909] env[61867]: DEBUG nova.compute.utils [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1035.656757] env[61867]: DEBUG nova.compute.manager [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1035.656926] env[61867]: DEBUG nova.network.neutron [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1035.659125] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c4dc336-e222-4794-92bc-aa3036c357b6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.677282] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Updating instance '986ada58-eb26-4289-bc0e-338c6a399645' progress to 67 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1035.701976] env[61867]: DEBUG nova.policy [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3d3b80154d0145bdb98e0afd8dd38d32', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cedf44d3b54a42798bb06ad6aae012e0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 1035.709957] env[61867]: DEBUG nova.compute.utils [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1035.758368] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48ba4664-1a62-4e2b-b93f-5a6fdd94241b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.765892] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b2da51-3311-4ea3-814b-b8d218b0be8f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.794802] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09a70f8c-721e-404a-af71-2780a96191d6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.801435] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30b93b16-34c0-4758-8142-324aaaf93734 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.813811] env[61867]: DEBUG nova.compute.provider_tree [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.939023] env[61867]: DEBUG nova.network.neutron [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Successfully created port: ddcbc24f-ddfe-46d9-8ad4-214cd357022d {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1036.157690] env[61867]: DEBUG nova.compute.manager [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1036.212624] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.225402] env[61867]: DEBUG nova.network.neutron [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Port 072ec5db-d2ce-4e94-809d-7ff5fd034562 binding to destination host cpu-1 is already ACTIVE {{(pid=61867) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1036.317190] env[61867]: DEBUG nova.scheduler.client.report [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1036.821641] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.679s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.824960] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.017s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.824960] env[61867]: DEBUG nova.objects.instance [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Lazy-loading 'resources' on Instance uuid 64947dbe-062c-4041-9ba9-d055279ded1e {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.844487] env[61867]: INFO nova.scheduler.client.report [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Deleted allocations for instance 4d7bfb31-d565-49e4-8c51-0122acebba2d [ 1037.169822] env[61867]: DEBUG nova.compute.manager [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1037.195817] env[61867]: DEBUG nova.virt.hardware [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1037.196075] env[61867]: DEBUG nova.virt.hardware [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1037.196269] env[61867]: DEBUG nova.virt.hardware [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1037.196475] env[61867]: DEBUG nova.virt.hardware [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1037.196639] env[61867]: DEBUG nova.virt.hardware [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1037.196785] env[61867]: DEBUG nova.virt.hardware [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1037.196996] env[61867]: DEBUG nova.virt.hardware [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1037.197177] env[61867]: DEBUG nova.virt.hardware [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1037.197347] env[61867]: DEBUG nova.virt.hardware [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1037.197510] env[61867]: DEBUG nova.virt.hardware [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1037.197732] env[61867]: DEBUG nova.virt.hardware [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1037.198654] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2690dab-3f4e-417f-b8c4-5df9983f8ac0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.207008] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99df812d-7370-48b7-ba65-a5ef5adb5503 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.246671] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "986ada58-eb26-4289-bc0e-338c6a399645-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.246671] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "986ada58-eb26-4289-bc0e-338c6a399645-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.246806] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "986ada58-eb26-4289-bc0e-338c6a399645-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.287968] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.288360] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.288478] env[61867]: INFO nova.compute.manager [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Attaching volume e8d5686e-270a-4ffc-9ba7-aa3f11f81510 to /dev/sdb [ 1037.319580] env[61867]: DEBUG nova.compute.manager [req-202926d6-5939-4367-94d3-9e5cc81992f0 req-ac0d5db4-af7f-460c-b971-4f224acda915 service nova] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Received event network-vif-plugged-ddcbc24f-ddfe-46d9-8ad4-214cd357022d {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.319801] env[61867]: DEBUG oslo_concurrency.lockutils [req-202926d6-5939-4367-94d3-9e5cc81992f0 req-ac0d5db4-af7f-460c-b971-4f224acda915 service nova] Acquiring lock "a9724422-abd1-4771-acb2-2c0ac45f5c16-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.320021] env[61867]: DEBUG oslo_concurrency.lockutils [req-202926d6-5939-4367-94d3-9e5cc81992f0 req-ac0d5db4-af7f-460c-b971-4f224acda915 service nova] Lock "a9724422-abd1-4771-acb2-2c0ac45f5c16-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.320191] env[61867]: DEBUG oslo_concurrency.lockutils [req-202926d6-5939-4367-94d3-9e5cc81992f0 req-ac0d5db4-af7f-460c-b971-4f224acda915 service nova] Lock "a9724422-abd1-4771-acb2-2c0ac45f5c16-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.320360] env[61867]: DEBUG nova.compute.manager [req-202926d6-5939-4367-94d3-9e5cc81992f0 req-ac0d5db4-af7f-460c-b971-4f224acda915 service nova] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] No waiting events found dispatching network-vif-plugged-ddcbc24f-ddfe-46d9-8ad4-214cd357022d {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1037.320525] env[61867]: WARNING nova.compute.manager [req-202926d6-5939-4367-94d3-9e5cc81992f0 req-ac0d5db4-af7f-460c-b971-4f224acda915 service nova] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Received unexpected event network-vif-plugged-ddcbc24f-ddfe-46d9-8ad4-214cd357022d for instance with vm_state building and task_state spawning. [ 1037.322613] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-316f3fe1-8533-43ad-a093-f0fff824d3c0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.334319] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f227fb7-1942-449f-bfc8-029251559960 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.346975] env[61867]: DEBUG nova.virt.block_device [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating existing volume attachment record: 38338bff-6427-47dd-bb65-f0e74635c748 {{(pid=61867) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1037.355148] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6072ac4d-8aeb-42c3-a285-c58dcbfeb44d tempest-AttachInterfacesTestJSON-845134924 tempest-AttachInterfacesTestJSON-845134924-project-member] Lock "4d7bfb31-d565-49e4-8c51-0122acebba2d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.526s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.407516] env[61867]: DEBUG nova.network.neutron [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Successfully updated port: ddcbc24f-ddfe-46d9-8ad4-214cd357022d {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1037.433732] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d993cd0-7d7e-4a20-964f-edce5246dfe7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.442112] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e55f4f-7daa-4a02-bd83-bc1acc13f74f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.471063] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c8f700e-cfd1-4d83-97ff-1aed298674d3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.477691] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519e50db-cd1c-4b21-8a7a-f699183a71bb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.490324] env[61867]: DEBUG nova.compute.provider_tree [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.910051] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "refresh_cache-a9724422-abd1-4771-acb2-2c0ac45f5c16" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.910051] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquired lock "refresh_cache-a9724422-abd1-4771-acb2-2c0ac45f5c16" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.910051] env[61867]: DEBUG nova.network.neutron [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1037.994226] env[61867]: DEBUG nova.scheduler.client.report [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1038.283726] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "refresh_cache-986ada58-eb26-4289-bc0e-338c6a399645" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.284032] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "refresh_cache-986ada58-eb26-4289-bc0e-338c6a399645" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.284177] env[61867]: DEBUG nova.network.neutron [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1038.442274] env[61867]: DEBUG nova.network.neutron [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1038.499669] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.676s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.522796] env[61867]: INFO nova.scheduler.client.report [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Deleted allocations for instance 64947dbe-062c-4041-9ba9-d055279ded1e [ 1038.562658] env[61867]: DEBUG nova.network.neutron [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Updating instance_info_cache with network_info: [{"id": "ddcbc24f-ddfe-46d9-8ad4-214cd357022d", "address": "fa:16:3e:42:f1:8f", "network": {"id": "116eb6c3-714d-4c39-a1dc-3d18b6b8d97e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-751125907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "cedf44d3b54a42798bb06ad6aae012e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddcbc24f-dd", "ovs_interfaceid": "ddcbc24f-ddfe-46d9-8ad4-214cd357022d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.976515] env[61867]: DEBUG nova.network.neutron [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Updating instance_info_cache with network_info: [{"id": "072ec5db-d2ce-4e94-809d-7ff5fd034562", "address": "fa:16:3e:45:de:03", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap072ec5db-d2", "ovs_interfaceid": "072ec5db-d2ce-4e94-809d-7ff5fd034562", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.033898] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d7e4b79b-9832-488f-ae35-f5b808f610fa tempest-ServerShowV257Test-479796155 tempest-ServerShowV257Test-479796155-project-member] Lock "64947dbe-062c-4041-9ba9-d055279ded1e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.922s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.066414] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Releasing lock "refresh_cache-a9724422-abd1-4771-acb2-2c0ac45f5c16" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.066679] env[61867]: DEBUG nova.compute.manager [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Instance network_info: |[{"id": "ddcbc24f-ddfe-46d9-8ad4-214cd357022d", "address": "fa:16:3e:42:f1:8f", "network": {"id": "116eb6c3-714d-4c39-a1dc-3d18b6b8d97e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-751125907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "cedf44d3b54a42798bb06ad6aae012e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddcbc24f-dd", "ovs_interfaceid": "ddcbc24f-ddfe-46d9-8ad4-214cd357022d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1039.067113] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:f1:8f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '535b175f-71d3-4226-81ab-ca253f27fedd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ddcbc24f-ddfe-46d9-8ad4-214cd357022d', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1039.077814] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Creating folder: Project (cedf44d3b54a42798bb06ad6aae012e0). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1039.078366] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5fbabcd3-821b-4b28-b45e-df427eccf3c7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.088508] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Created folder: Project (cedf44d3b54a42798bb06ad6aae012e0) in parent group-v274258. [ 1039.088704] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Creating folder: Instances. Parent ref: group-v274422. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1039.088936] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ffe427ab-d59f-431e-b4d1-d6de9cc67d49 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.097484] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Created folder: Instances in parent group-v274422. [ 1039.097733] env[61867]: DEBUG oslo.service.loopingcall [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.097928] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1039.098152] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b12b3c7a-ff9b-471c-914f-3d093f1926df {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.119260] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1039.119260] env[61867]: value = "task-1277116" [ 1039.119260] env[61867]: _type = "Task" [ 1039.119260] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.131044] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277116, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.351458] env[61867]: DEBUG nova.compute.manager [req-27238a94-833f-4b63-befc-906ad369cbec req-c52feb87-e0ca-48fd-a384-c73f4a1138bc service nova] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Received event network-changed-ddcbc24f-ddfe-46d9-8ad4-214cd357022d {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.351690] env[61867]: DEBUG nova.compute.manager [req-27238a94-833f-4b63-befc-906ad369cbec req-c52feb87-e0ca-48fd-a384-c73f4a1138bc service nova] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Refreshing instance network info cache due to event network-changed-ddcbc24f-ddfe-46d9-8ad4-214cd357022d. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1039.352374] env[61867]: DEBUG oslo_concurrency.lockutils [req-27238a94-833f-4b63-befc-906ad369cbec req-c52feb87-e0ca-48fd-a384-c73f4a1138bc service nova] Acquiring lock "refresh_cache-a9724422-abd1-4771-acb2-2c0ac45f5c16" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.352374] env[61867]: DEBUG oslo_concurrency.lockutils [req-27238a94-833f-4b63-befc-906ad369cbec req-c52feb87-e0ca-48fd-a384-c73f4a1138bc service nova] Acquired lock "refresh_cache-a9724422-abd1-4771-acb2-2c0ac45f5c16" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.352374] env[61867]: DEBUG nova.network.neutron [req-27238a94-833f-4b63-befc-906ad369cbec req-c52feb87-e0ca-48fd-a384-c73f4a1138bc service nova] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Refreshing network info cache for port ddcbc24f-ddfe-46d9-8ad4-214cd357022d {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1039.478557] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "refresh_cache-986ada58-eb26-4289-bc0e-338c6a399645" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.629201] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277116, 'name': CreateVM_Task, 'duration_secs': 0.304373} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.630038] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1039.630158] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.630365] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.630728] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1039.631935] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ae65deb-0670-4ca3-b3a5-4fd7fa7693ab {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.635696] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1039.635696] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520e8b17-9905-9a40-fa11-6650cb35bf7d" [ 1039.635696] env[61867]: _type = "Task" [ 1039.635696] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.643383] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520e8b17-9905-9a40-fa11-6650cb35bf7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.003752] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf67032-c714-4a7f-b730-a5308f4dd90a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.023820] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0cd1aa-d45e-450e-aebf-027ec2b65d22 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.030340] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Updating instance '986ada58-eb26-4289-bc0e-338c6a399645' progress to 83 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1040.058396] env[61867]: DEBUG nova.network.neutron [req-27238a94-833f-4b63-befc-906ad369cbec req-c52feb87-e0ca-48fd-a384-c73f4a1138bc service nova] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Updated VIF entry in instance network info cache for port ddcbc24f-ddfe-46d9-8ad4-214cd357022d. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1040.058757] env[61867]: DEBUG nova.network.neutron [req-27238a94-833f-4b63-befc-906ad369cbec req-c52feb87-e0ca-48fd-a384-c73f4a1138bc service nova] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Updating instance_info_cache with network_info: [{"id": "ddcbc24f-ddfe-46d9-8ad4-214cd357022d", "address": "fa:16:3e:42:f1:8f", "network": {"id": "116eb6c3-714d-4c39-a1dc-3d18b6b8d97e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-751125907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "cedf44d3b54a42798bb06ad6aae012e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddcbc24f-dd", "ovs_interfaceid": "ddcbc24f-ddfe-46d9-8ad4-214cd357022d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.147246] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520e8b17-9905-9a40-fa11-6650cb35bf7d, 'name': SearchDatastore_Task, 'duration_secs': 0.038612} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.147350] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.147575] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1040.147814] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.147966] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.148177] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1040.148448] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-94e8079a-510b-4927-a83c-d55780303f32 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.156382] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1040.156566] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1040.157264] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e176043-3b78-4c6d-b256-540577b436cf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.162044] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1040.162044] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e7b7c0-588e-ab16-0e35-21bdce4b8d1f" [ 1040.162044] env[61867]: _type = "Task" [ 1040.162044] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.169554] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e7b7c0-588e-ab16-0e35-21bdce4b8d1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.536913] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1040.537303] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a1bbe59-95d5-4c18-9d8f-f4a8c9a708ae {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.544150] env[61867]: DEBUG oslo_vmware.api [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 1040.544150] env[61867]: value = "task-1277118" [ 1040.544150] env[61867]: _type = "Task" [ 1040.544150] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.552093] env[61867]: DEBUG oslo_vmware.api [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277118, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.561706] env[61867]: DEBUG oslo_concurrency.lockutils [req-27238a94-833f-4b63-befc-906ad369cbec req-c52feb87-e0ca-48fd-a384-c73f4a1138bc service nova] Releasing lock "refresh_cache-a9724422-abd1-4771-acb2-2c0ac45f5c16" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.673123] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e7b7c0-588e-ab16-0e35-21bdce4b8d1f, 'name': SearchDatastore_Task, 'duration_secs': 0.008532} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.673347] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be6e0026-9230-4e0f-ac26-305353ffd99b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.678236] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1040.678236] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f3105e-dd2c-89d7-2a79-7b6609069198" [ 1040.678236] env[61867]: _type = "Task" [ 1040.678236] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.685574] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f3105e-dd2c-89d7-2a79-7b6609069198, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.053391] env[61867]: DEBUG oslo_vmware.api [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277118, 'name': PowerOnVM_Task, 'duration_secs': 0.387304} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.054098] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1041.054098] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9343afde-36e7-4303-8b59-eb7fdd40e409 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Updating instance '986ada58-eb26-4289-bc0e-338c6a399645' progress to 100 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1041.188500] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f3105e-dd2c-89d7-2a79-7b6609069198, 'name': SearchDatastore_Task, 'duration_secs': 0.008958} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.188760] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.189032] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] a9724422-abd1-4771-acb2-2c0ac45f5c16/a9724422-abd1-4771-acb2-2c0ac45f5c16.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1041.189289] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-70463316-168e-4119-8669-d186e1c9060c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.195284] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1041.195284] env[61867]: value = "task-1277119" [ 1041.195284] env[61867]: _type = "Task" [ 1041.195284] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.202912] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277119, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.706840] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277119, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.898908] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Volume attach. Driver type: vmdk {{(pid=61867) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1041.899175] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274421', 'volume_id': 'e8d5686e-270a-4ffc-9ba7-aa3f11f81510', 'name': 'volume-e8d5686e-270a-4ffc-9ba7-aa3f11f81510', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2', 'attached_at': '', 'detached_at': '', 'volume_id': 'e8d5686e-270a-4ffc-9ba7-aa3f11f81510', 'serial': 'e8d5686e-270a-4ffc-9ba7-aa3f11f81510'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1041.900086] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0bcda48-0e7e-4e1f-b6a1-7a160f185106 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.925312] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b713907e-27f7-4d22-ab10-a58f8ca2867c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.963329] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] volume-e8d5686e-270a-4ffc-9ba7-aa3f11f81510/volume-e8d5686e-270a-4ffc-9ba7-aa3f11f81510.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1041.963739] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7733070-bca9-40d1-8423-aea6ed443f41 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.990189] env[61867]: DEBUG oslo_vmware.api [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1041.990189] env[61867]: value = "task-1277120" [ 1041.990189] env[61867]: _type = "Task" [ 1041.990189] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.000898] env[61867]: DEBUG oslo_vmware.api [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277120, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.206959] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277119, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.630114} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.207214] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] a9724422-abd1-4771-acb2-2c0ac45f5c16/a9724422-abd1-4771-acb2-2c0ac45f5c16.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1042.207437] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1042.207684] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a251085-2e28-484c-a613-24fd61d063e6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.214338] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1042.214338] env[61867]: value = "task-1277121" [ 1042.214338] env[61867]: _type = "Task" [ 1042.214338] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.221895] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277121, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.499887] env[61867]: DEBUG oslo_vmware.api [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277120, 'name': ReconfigVM_Task, 'duration_secs': 0.33067} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.499887] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Reconfigured VM instance instance-00000064 to attach disk [datastore2] volume-e8d5686e-270a-4ffc-9ba7-aa3f11f81510/volume-e8d5686e-270a-4ffc-9ba7-aa3f11f81510.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1042.504353] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07afa91c-27a7-457f-b5bf-4be9357b4f24 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.518880] env[61867]: DEBUG oslo_vmware.api [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1042.518880] env[61867]: value = "task-1277122" [ 1042.518880] env[61867]: _type = "Task" [ 1042.518880] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.528978] env[61867]: DEBUG oslo_vmware.api [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277122, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.724526] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277121, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065328} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.724882] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1042.725557] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45bf66bb-f018-4976-8d30-3a7a5e5dfcf8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.747030] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] a9724422-abd1-4771-acb2-2c0ac45f5c16/a9724422-abd1-4771-acb2-2c0ac45f5c16.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1042.747295] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-937347c1-617a-4fda-8dda-9a96812f3815 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.765957] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1042.765957] env[61867]: value = "task-1277123" [ 1042.765957] env[61867]: _type = "Task" [ 1042.765957] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.774728] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277123, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.029957] env[61867]: DEBUG oslo_vmware.api [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277122, 'name': ReconfigVM_Task, 'duration_secs': 0.15178} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.030393] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274421', 'volume_id': 'e8d5686e-270a-4ffc-9ba7-aa3f11f81510', 'name': 'volume-e8d5686e-270a-4ffc-9ba7-aa3f11f81510', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2', 'attached_at': '', 'detached_at': '', 'volume_id': 'e8d5686e-270a-4ffc-9ba7-aa3f11f81510', 'serial': 'e8d5686e-270a-4ffc-9ba7-aa3f11f81510'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1043.097090] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "986ada58-eb26-4289-bc0e-338c6a399645" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.097292] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "986ada58-eb26-4289-bc0e-338c6a399645" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.097532] env[61867]: DEBUG nova.compute.manager [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Going to confirm migration 5 {{(pid=61867) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1043.275798] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277123, 'name': ReconfigVM_Task, 'duration_secs': 0.244854} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.276116] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Reconfigured VM instance instance-00000069 to attach disk [datastore2] a9724422-abd1-4771-acb2-2c0ac45f5c16/a9724422-abd1-4771-acb2-2c0ac45f5c16.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1043.276760] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-582292e5-d357-4d15-b8de-d4d8c4c6dc20 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.283435] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1043.283435] env[61867]: value = "task-1277124" [ 1043.283435] env[61867]: _type = "Task" [ 1043.283435] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.290993] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277124, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.658036] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "refresh_cache-986ada58-eb26-4289-bc0e-338c6a399645" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.658252] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquired lock "refresh_cache-986ada58-eb26-4289-bc0e-338c6a399645" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.658434] env[61867]: DEBUG nova.network.neutron [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1043.658624] env[61867]: DEBUG nova.objects.instance [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lazy-loading 'info_cache' on Instance uuid 986ada58-eb26-4289-bc0e-338c6a399645 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.793319] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277124, 'name': Rename_Task, 'duration_secs': 0.130401} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.793655] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1043.793782] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c447c27b-4a4c-40e3-ae9d-e6cf25b4c7d4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.800352] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1043.800352] env[61867]: value = "task-1277125" [ 1043.800352] env[61867]: _type = "Task" [ 1043.800352] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.807361] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277125, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.064680] env[61867]: DEBUG nova.objects.instance [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lazy-loading 'flavor' on Instance uuid df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.311387] env[61867]: DEBUG oslo_vmware.api [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277125, 'name': PowerOnVM_Task, 'duration_secs': 0.388008} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.311758] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1044.312094] env[61867]: INFO nova.compute.manager [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Took 7.14 seconds to spawn the instance on the hypervisor. [ 1044.312242] env[61867]: DEBUG nova.compute.manager [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1044.313266] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ae65c6-2df2-49b8-865a-33fc103119f0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.570622] env[61867]: DEBUG oslo_concurrency.lockutils [None req-4f60120a-4661-417f-acd5-0b17c047ccc3 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.282s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.830614] env[61867]: INFO nova.compute.manager [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Took 11.89 seconds to build instance. [ 1044.847168] env[61867]: DEBUG nova.network.neutron [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Updating instance_info_cache with network_info: [{"id": "072ec5db-d2ce-4e94-809d-7ff5fd034562", "address": "fa:16:3e:45:de:03", "network": {"id": "12ffb1fc-0aef-496c-8cc2-f8166cfbbae8", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1221793896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28a8e133444440eb83dc6a848ef591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap072ec5db-d2", "ovs_interfaceid": "072ec5db-d2ce-4e94-809d-7ff5fd034562", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.144097] env[61867]: INFO nova.compute.manager [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Rescuing [ 1045.144362] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "refresh_cache-a9724422-abd1-4771-acb2-2c0ac45f5c16" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.144525] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquired lock "refresh_cache-a9724422-abd1-4771-acb2-2c0ac45f5c16" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.144701] env[61867]: DEBUG nova.network.neutron [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1045.331917] env[61867]: DEBUG oslo_concurrency.lockutils [None req-b4af32f6-e234-4442-84d3-0ea93b4b627e tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "a9724422-abd1-4771-acb2-2c0ac45f5c16" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.400s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.350037] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Releasing lock "refresh_cache-986ada58-eb26-4289-bc0e-338c6a399645" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.350218] env[61867]: DEBUG nova.objects.instance [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lazy-loading 'migration_context' on Instance uuid 986ada58-eb26-4289-bc0e-338c6a399645 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.366882] env[61867]: DEBUG nova.compute.manager [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Stashing vm_state: active {{(pid=61867) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1045.814550] env[61867]: DEBUG nova.network.neutron [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Updating instance_info_cache with network_info: [{"id": "ddcbc24f-ddfe-46d9-8ad4-214cd357022d", "address": "fa:16:3e:42:f1:8f", "network": {"id": "116eb6c3-714d-4c39-a1dc-3d18b6b8d97e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-751125907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "cedf44d3b54a42798bb06ad6aae012e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddcbc24f-dd", "ovs_interfaceid": "ddcbc24f-ddfe-46d9-8ad4-214cd357022d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.854065] env[61867]: DEBUG nova.objects.base [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Object Instance<986ada58-eb26-4289-bc0e-338c6a399645> lazy-loaded attributes: info_cache,migration_context {{(pid=61867) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1045.855010] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9a5edfc-6deb-4bfb-a2eb-3235c3252977 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.877868] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d892f18e-e8a0-4dc9-95ad-094537c41313 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.883360] env[61867]: DEBUG oslo_vmware.api [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 1045.883360] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527809c0-9497-6886-08e5-a3359005dd6b" [ 1045.883360] env[61867]: _type = "Task" [ 1045.883360] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.891850] env[61867]: DEBUG oslo_vmware.api [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527809c0-9497-6886-08e5-a3359005dd6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.892766] env[61867]: DEBUG oslo_concurrency.lockutils [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.892999] env[61867]: DEBUG oslo_concurrency.lockutils [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.317616] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Releasing lock "refresh_cache-a9724422-abd1-4771-acb2-2c0ac45f5c16" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.394248] env[61867]: DEBUG oslo_vmware.api [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527809c0-9497-6886-08e5-a3359005dd6b, 'name': SearchDatastore_Task, 'duration_secs': 0.008216} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.394560] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.396950] env[61867]: INFO nova.compute.claims [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1046.847340] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1046.847686] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb77ed33-0fd9-4686-aadb-9b8ab5c91414 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.855330] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1046.855330] env[61867]: value = "task-1277126" [ 1046.855330] env[61867]: _type = "Task" [ 1046.855330] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.865117] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277126, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.902827] env[61867]: INFO nova.compute.resource_tracker [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating resource usage from migration b428c0ea-e305-41b1-bae7-2d1c68bd7059 [ 1046.991087] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c64c78-c272-452a-979e-827ac1876ae1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.001360] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6df854-d76f-463e-938c-d34b01d0cc6b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.030478] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78aa1a10-306b-46d5-bba8-54f6e5d327ab {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.037523] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd95e8e-c850-43a2-8a03-3a9b6901bf12 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.050150] env[61867]: DEBUG nova.compute.provider_tree [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1047.365069] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277126, 'name': PowerOffVM_Task, 'duration_secs': 0.174676} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.365352] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1047.366115] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69968158-30f7-4c53-bc49-f9965b6eb929 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.384560] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66dd9ecf-dc32-488e-b227-265c64401fea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.411356] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1047.411638] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6184adab-9d49-493a-8fb8-394d308a49e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.419378] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1047.419378] env[61867]: value = "task-1277127" [ 1047.419378] env[61867]: _type = "Task" [ 1047.419378] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.428148] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] VM already powered off {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1047.428411] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1047.428650] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.428804] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.428993] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1047.429248] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64872447-8628-4335-aad2-f86973adf6cc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.444158] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1047.444369] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1047.445334] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55959967-6a7a-4808-9609-e845ca55b570 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.450283] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1047.450283] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52421296-e301-03f3-eadb-3208500efee4" [ 1047.450283] env[61867]: _type = "Task" [ 1047.450283] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.457361] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52421296-e301-03f3-eadb-3208500efee4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.553392] env[61867]: DEBUG nova.scheduler.client.report [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1047.961531] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52421296-e301-03f3-eadb-3208500efee4, 'name': SearchDatastore_Task, 'duration_secs': 0.027096} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.962491] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2847d3b1-c4a5-4f67-b302-995ad9c4bb0e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.967881] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1047.967881] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ccb949-7a2a-e2b0-2fc4-5155cc7a2736" [ 1047.967881] env[61867]: _type = "Task" [ 1047.967881] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.975555] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ccb949-7a2a-e2b0-2fc4-5155cc7a2736, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.057808] env[61867]: DEBUG oslo_concurrency.lockutils [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.165s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.058038] env[61867]: INFO nova.compute.manager [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Migrating [ 1048.064832] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.670s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.482772] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52ccb949-7a2a-e2b0-2fc4-5155cc7a2736, 'name': SearchDatastore_Task, 'duration_secs': 0.017936} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.483227] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.483682] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] a9724422-abd1-4771-acb2-2c0ac45f5c16/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk. {{(pid=61867) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1048.484105] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5f82fe60-7dcf-495e-a499-705b25852f51 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.492345] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1048.492345] env[61867]: value = "task-1277128" [ 1048.492345] env[61867]: _type = "Task" [ 1048.492345] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.503378] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277128, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.575679] env[61867]: DEBUG oslo_concurrency.lockutils [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.575679] env[61867]: DEBUG oslo_concurrency.lockutils [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.576123] env[61867]: DEBUG nova.network.neutron [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1048.650237] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8274280c-c369-4cfb-bb69-3ba6d3826d1e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.658658] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-390ecc48-58e5-4968-8872-b58cac8a15f7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.693167] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd61252-5e1c-4232-93a9-8ce36de70b3b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.701037] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e807f5a-7bcb-4ce7-9e8b-020be3dd1b65 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.714941] env[61867]: DEBUG nova.compute.provider_tree [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.001982] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277128, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455581} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.002351] env[61867]: INFO nova.virt.vmwareapi.ds_util [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] a9724422-abd1-4771-acb2-2c0ac45f5c16/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk. [ 1049.003034] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b77261-bae9-4739-952a-23273be18b16 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.026423] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] a9724422-abd1-4771-acb2-2c0ac45f5c16/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.026649] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75ee2923-6fe3-41f7-9923-90f271b65800 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.044815] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1049.044815] env[61867]: value = "task-1277129" [ 1049.044815] env[61867]: _type = "Task" [ 1049.044815] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.051790] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277129, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.218448] env[61867]: DEBUG nova.scheduler.client.report [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1049.289047] env[61867]: DEBUG nova.network.neutron [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating instance_info_cache with network_info: [{"id": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "address": "fa:16:3e:b9:62:3b", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1e84fbf-9f", "ovs_interfaceid": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.554642] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277129, 'name': ReconfigVM_Task, 'duration_secs': 0.281106} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.554883] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Reconfigured VM instance instance-00000069 to attach disk [datastore2] a9724422-abd1-4771-acb2-2c0ac45f5c16/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1049.555698] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0578dfdc-a7b2-4f3c-9fd6-450e9b0c8ad1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.581151] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7e8fa78-b1c3-4964-ae34-4c3985771bcc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.596627] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1049.596627] env[61867]: value = "task-1277130" [ 1049.596627] env[61867]: _type = "Task" [ 1049.596627] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.604163] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277130, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.791597] env[61867]: DEBUG oslo_concurrency.lockutils [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.106114] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277130, 'name': ReconfigVM_Task, 'duration_secs': 0.147154} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.106545] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1050.106637] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d38b162b-6777-4327-bafe-859cf5dd5f67 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.113891] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1050.113891] env[61867]: value = "task-1277131" [ 1050.113891] env[61867]: _type = "Task" [ 1050.113891] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.121382] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277131, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.228481] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.164s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.623793] env[61867]: DEBUG oslo_vmware.api [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277131, 'name': PowerOnVM_Task, 'duration_secs': 0.359968} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.624035] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1050.626761] env[61867]: DEBUG nova.compute.manager [None req-0933ac5b-c221-44c2-a62f-7656563a9799 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1050.627531] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249eeacc-44ee-4ba3-b8dd-497f80fbb773 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.779960] env[61867]: INFO nova.scheduler.client.report [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleted allocation for migration 2ffb5793-9433-4a7a-b35c-ee8968f54a67 [ 1051.287697] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "986ada58-eb26-4289-bc0e-338c6a399645" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.190s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.305387] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aae16dd-a591-4fe9-8bcd-86d965a5841c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.328115] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating instance 'df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2' progress to 0 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1051.833706] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1051.834095] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2b97ca52-4306-4b72-9239-9aed6fed7982 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.840785] env[61867]: DEBUG oslo_vmware.api [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1051.840785] env[61867]: value = "task-1277132" [ 1051.840785] env[61867]: _type = "Task" [ 1051.840785] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.851751] env[61867]: DEBUG oslo_vmware.api [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277132, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.319585] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "70170813-af35-4d7f-88e6-2adee247a701" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.319974] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "70170813-af35-4d7f-88e6-2adee247a701" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.349627] env[61867]: DEBUG oslo_vmware.api [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277132, 'name': PowerOffVM_Task, 'duration_secs': 0.230325} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.349878] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1052.350078] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating instance 'df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2' progress to 17 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1052.621381] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "986ada58-eb26-4289-bc0e-338c6a399645" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.621626] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "986ada58-eb26-4289-bc0e-338c6a399645" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.621848] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "986ada58-eb26-4289-bc0e-338c6a399645-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.622062] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "986ada58-eb26-4289-bc0e-338c6a399645-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.622234] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "986ada58-eb26-4289-bc0e-338c6a399645-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.624555] env[61867]: INFO nova.compute.manager [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Terminating instance [ 1052.626348] env[61867]: DEBUG nova.compute.manager [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1052.626571] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1052.627393] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5787453-271e-44e3-b063-4e4daa3ac757 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.634901] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1052.635155] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bb9d9c2e-1c46-4d5d-8a31-5e5e96b7c803 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.641220] env[61867]: DEBUG oslo_vmware.api [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 1052.641220] env[61867]: value = "task-1277133" [ 1052.641220] env[61867]: _type = "Task" [ 1052.641220] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.648814] env[61867]: DEBUG oslo_vmware.api [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277133, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.821920] env[61867]: DEBUG nova.compute.manager [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1052.856104] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1052.856359] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1052.856527] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1052.856710] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1052.856862] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1052.857022] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1052.857237] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1052.857400] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1052.857569] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1052.857732] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1052.857905] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1052.863930] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84779b14-243c-446c-8371-5b327f120342 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.880426] env[61867]: DEBUG oslo_vmware.api [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1052.880426] env[61867]: value = "task-1277134" [ 1052.880426] env[61867]: _type = "Task" [ 1052.880426] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.888600] env[61867]: DEBUG oslo_vmware.api [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277134, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.150637] env[61867]: DEBUG oslo_vmware.api [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277133, 'name': PowerOffVM_Task, 'duration_secs': 0.176104} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.150904] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1053.151094] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1053.151351] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-52095f12-ce41-47a3-9f57-34103f9b6253 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.207662] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1053.207980] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1053.208193] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleting the datastore file [datastore2] 986ada58-eb26-4289-bc0e-338c6a399645 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1053.208519] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b70256a-ebca-4c3a-a7a7-9a214795dcb7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.214677] env[61867]: DEBUG oslo_vmware.api [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for the task: (returnval){ [ 1053.214677] env[61867]: value = "task-1277136" [ 1053.214677] env[61867]: _type = "Task" [ 1053.214677] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.224431] env[61867]: DEBUG oslo_vmware.api [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277136, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.350143] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.350410] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.351978] env[61867]: INFO nova.compute.claims [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1053.390777] env[61867]: DEBUG oslo_vmware.api [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277134, 'name': ReconfigVM_Task, 'duration_secs': 0.152511} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.391078] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating instance 'df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2' progress to 33 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1053.724928] env[61867]: DEBUG oslo_vmware.api [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Task: {'id': task-1277136, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.279973} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.725152] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1053.725341] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1053.725525] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1053.725705] env[61867]: INFO nova.compute.manager [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1053.725951] env[61867]: DEBUG oslo.service.loopingcall [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.726165] env[61867]: DEBUG nova.compute.manager [-] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1053.726261] env[61867]: DEBUG nova.network.neutron [-] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1053.897403] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1053.897716] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1053.897933] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1053.898138] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1053.898295] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1053.898447] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1053.898661] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1053.898809] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1053.898973] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1053.899192] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1053.899388] env[61867]: DEBUG nova.virt.hardware [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1053.904670] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Reconfiguring VM instance instance-00000064 to detach disk 2000 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1053.905281] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80be68da-5046-480d-9e7a-c7aed6785a6a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.924835] env[61867]: DEBUG oslo_vmware.api [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1053.924835] env[61867]: value = "task-1277137" [ 1053.924835] env[61867]: _type = "Task" [ 1053.924835] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.933206] env[61867]: DEBUG oslo_vmware.api [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277137, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.995404] env[61867]: DEBUG nova.compute.manager [req-ee84c202-9cd0-4b3f-8a0f-d1a8945f5b32 req-91c5419b-abcf-491a-bc4f-6e9f2b8b6d64 service nova] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Received event network-vif-deleted-072ec5db-d2ce-4e94-809d-7ff5fd034562 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1053.995579] env[61867]: INFO nova.compute.manager [req-ee84c202-9cd0-4b3f-8a0f-d1a8945f5b32 req-91c5419b-abcf-491a-bc4f-6e9f2b8b6d64 service nova] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Neutron deleted interface 072ec5db-d2ce-4e94-809d-7ff5fd034562; detaching it from the instance and deleting it from the info cache [ 1053.995762] env[61867]: DEBUG nova.network.neutron [req-ee84c202-9cd0-4b3f-8a0f-d1a8945f5b32 req-91c5419b-abcf-491a-bc4f-6e9f2b8b6d64 service nova] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.435660] env[61867]: DEBUG oslo_vmware.api [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277137, 'name': ReconfigVM_Task, 'duration_secs': 0.272294} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.435953] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Reconfigured VM instance instance-00000064 to detach disk 2000 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1054.436739] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceab753b-47ba-40ab-b3ec-5a6c1174ab0e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.441672] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-699b30be-f985-4cce-9548-2c346d507905 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.463596] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2/df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1054.464173] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ba3691a-f67b-4abd-a5cc-f12347d38621 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.478669] env[61867]: DEBUG nova.network.neutron [-] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.480854] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8021d17-a0fe-43d1-9b1e-255676c4a4cb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.485025] env[61867]: DEBUG oslo_vmware.api [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1054.485025] env[61867]: value = "task-1277138" [ 1054.485025] env[61867]: _type = "Task" [ 1054.485025] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.517710] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-20ead4d5-0d05-4db4-a69c-0e05de7881b2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.520197] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7c976c-1c09-4a57-ac4c-f4e6ca1a3f03 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.525680] env[61867]: DEBUG oslo_vmware.api [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277138, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.531375] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b360f2f-7867-4a86-bbeb-f0b1702c096e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.537430] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d8a08f6-c1c7-4a7f-9d43-26b8f6e85add {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.556864] env[61867]: DEBUG nova.compute.provider_tree [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1054.563223] env[61867]: DEBUG nova.compute.manager [req-ee84c202-9cd0-4b3f-8a0f-d1a8945f5b32 req-91c5419b-abcf-491a-bc4f-6e9f2b8b6d64 service nova] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Detach interface failed, port_id=072ec5db-d2ce-4e94-809d-7ff5fd034562, reason: Instance 986ada58-eb26-4289-bc0e-338c6a399645 could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1054.985312] env[61867]: INFO nova.compute.manager [-] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Took 1.26 seconds to deallocate network for instance. [ 1054.999755] env[61867]: DEBUG oslo_vmware.api [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277138, 'name': ReconfigVM_Task, 'duration_secs': 0.274515} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.999985] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Reconfigured VM instance instance-00000064 to attach disk [datastore1] df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2/df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1055.000293] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating instance 'df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2' progress to 50 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1055.064492] env[61867]: DEBUG nova.scheduler.client.report [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1055.496185] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.506490] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc529594-9e57-4627-a11b-28ee6f096193 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.527738] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ea7c32-5325-426b-8bf3-bfa1230cbf04 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.547265] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating instance 'df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2' progress to 67 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1055.571644] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.221s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.572160] env[61867]: DEBUG nova.compute.manager [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1055.574518] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.079s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.574725] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.596430] env[61867]: INFO nova.scheduler.client.report [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Deleted allocations for instance 986ada58-eb26-4289-bc0e-338c6a399645 [ 1055.850608] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "748e8a71-28b6-4c91-9cbb-856359348c96" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.850865] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "748e8a71-28b6-4c91-9cbb-856359348c96" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.078690] env[61867]: DEBUG nova.compute.utils [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1056.080067] env[61867]: DEBUG nova.compute.manager [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1056.080239] env[61867]: DEBUG nova.network.neutron [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1056.106035] env[61867]: DEBUG oslo_concurrency.lockutils [None req-303b2d3d-079c-4d3e-9849-017600740db7 tempest-DeleteServersTestJSON-559044558 tempest-DeleteServersTestJSON-559044558-project-member] Lock "986ada58-eb26-4289-bc0e-338c6a399645" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.484s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.121777] env[61867]: DEBUG nova.policy [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3d3b80154d0145bdb98e0afd8dd38d32', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cedf44d3b54a42798bb06ad6aae012e0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 1056.354117] env[61867]: DEBUG nova.compute.utils [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1056.364332] env[61867]: DEBUG nova.network.neutron [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Successfully created port: b7b25937-4b43-48af-88a2-628f656bbf4d {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1056.582912] env[61867]: DEBUG nova.compute.manager [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1056.856890] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "748e8a71-28b6-4c91-9cbb-856359348c96" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.191361] env[61867]: DEBUG nova.network.neutron [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Port f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2 binding to destination host cpu-1 is already ACTIVE {{(pid=61867) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1057.594868] env[61867]: DEBUG nova.compute.manager [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1057.619930] env[61867]: DEBUG nova.virt.hardware [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1057.620219] env[61867]: DEBUG nova.virt.hardware [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1057.620385] env[61867]: DEBUG nova.virt.hardware [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1057.620569] env[61867]: DEBUG nova.virt.hardware [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1057.620728] env[61867]: DEBUG nova.virt.hardware [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1057.620888] env[61867]: DEBUG nova.virt.hardware [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1057.621143] env[61867]: DEBUG nova.virt.hardware [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1057.621336] env[61867]: DEBUG nova.virt.hardware [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1057.621514] env[61867]: DEBUG nova.virt.hardware [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1057.621682] env[61867]: DEBUG nova.virt.hardware [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1057.621858] env[61867]: DEBUG nova.virt.hardware [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1057.622754] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13cc3063-210f-4252-aa6f-2fc4b50e1892 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.630993] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a1d90a2-5fc6-4618-8b9b-88b8723ab784 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.734311] env[61867]: DEBUG nova.compute.manager [req-041088c6-1695-4739-ac2b-edcf999dead8 req-81e29c8c-d2c6-4f4d-b0c8-18f173129615 service nova] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Received event network-vif-plugged-b7b25937-4b43-48af-88a2-628f656bbf4d {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1057.734638] env[61867]: DEBUG oslo_concurrency.lockutils [req-041088c6-1695-4739-ac2b-edcf999dead8 req-81e29c8c-d2c6-4f4d-b0c8-18f173129615 service nova] Acquiring lock "70170813-af35-4d7f-88e6-2adee247a701-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.734851] env[61867]: DEBUG oslo_concurrency.lockutils [req-041088c6-1695-4739-ac2b-edcf999dead8 req-81e29c8c-d2c6-4f4d-b0c8-18f173129615 service nova] Lock "70170813-af35-4d7f-88e6-2adee247a701-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.735913] env[61867]: DEBUG oslo_concurrency.lockutils [req-041088c6-1695-4739-ac2b-edcf999dead8 req-81e29c8c-d2c6-4f4d-b0c8-18f173129615 service nova] Lock "70170813-af35-4d7f-88e6-2adee247a701-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.735913] env[61867]: DEBUG nova.compute.manager [req-041088c6-1695-4739-ac2b-edcf999dead8 req-81e29c8c-d2c6-4f4d-b0c8-18f173129615 service nova] [instance: 70170813-af35-4d7f-88e6-2adee247a701] No waiting events found dispatching network-vif-plugged-b7b25937-4b43-48af-88a2-628f656bbf4d {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1057.735913] env[61867]: WARNING nova.compute.manager [req-041088c6-1695-4739-ac2b-edcf999dead8 req-81e29c8c-d2c6-4f4d-b0c8-18f173129615 service nova] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Received unexpected event network-vif-plugged-b7b25937-4b43-48af-88a2-628f656bbf4d for instance with vm_state building and task_state spawning. [ 1057.826122] env[61867]: DEBUG nova.network.neutron [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Successfully updated port: b7b25937-4b43-48af-88a2-628f656bbf4d {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1057.914743] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "748e8a71-28b6-4c91-9cbb-856359348c96" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.915024] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "748e8a71-28b6-4c91-9cbb-856359348c96" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.915303] env[61867]: INFO nova.compute.manager [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Attaching volume ae2bf781-d644-4860-a51b-9a8e3602b5fd to /dev/sdb [ 1057.945484] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c62e77f-8a4c-40d3-8a10-616067e00854 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.952725] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8dd7bca-485a-473f-aaa7-8bda78525cb0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.965132] env[61867]: DEBUG nova.virt.block_device [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Updating existing volume attachment record: bffd4767-263f-44b7-bf88-6b7c99e3abcf {{(pid=61867) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1058.215432] env[61867]: DEBUG oslo_concurrency.lockutils [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.215708] env[61867]: DEBUG oslo_concurrency.lockutils [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.215902] env[61867]: DEBUG oslo_concurrency.lockutils [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.329123] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "refresh_cache-70170813-af35-4d7f-88e6-2adee247a701" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.329273] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquired lock "refresh_cache-70170813-af35-4d7f-88e6-2adee247a701" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.329405] env[61867]: DEBUG nova.network.neutron [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1058.860146] env[61867]: DEBUG nova.network.neutron [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1058.957329] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1058.957488] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Cleaning up deleted instances with incomplete migration {{(pid=61867) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1058.987342] env[61867]: DEBUG nova.network.neutron [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Updating instance_info_cache with network_info: [{"id": "b7b25937-4b43-48af-88a2-628f656bbf4d", "address": "fa:16:3e:df:51:fa", "network": {"id": "116eb6c3-714d-4c39-a1dc-3d18b6b8d97e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-751125907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "cedf44d3b54a42798bb06ad6aae012e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7b25937-4b", "ovs_interfaceid": "b7b25937-4b43-48af-88a2-628f656bbf4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.245906] env[61867]: DEBUG oslo_concurrency.lockutils [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.246129] env[61867]: DEBUG oslo_concurrency.lockutils [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.246314] env[61867]: DEBUG nova.network.neutron [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1059.490573] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Releasing lock "refresh_cache-70170813-af35-4d7f-88e6-2adee247a701" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.491299] env[61867]: DEBUG nova.compute.manager [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Instance network_info: |[{"id": "b7b25937-4b43-48af-88a2-628f656bbf4d", "address": "fa:16:3e:df:51:fa", "network": {"id": "116eb6c3-714d-4c39-a1dc-3d18b6b8d97e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-751125907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "cedf44d3b54a42798bb06ad6aae012e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7b25937-4b", "ovs_interfaceid": "b7b25937-4b43-48af-88a2-628f656bbf4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1059.491451] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:51:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '535b175f-71d3-4226-81ab-ca253f27fedd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b7b25937-4b43-48af-88a2-628f656bbf4d', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1059.498941] env[61867]: DEBUG oslo.service.loopingcall [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1059.499176] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1059.499415] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-23a08269-268a-4e0b-bb1d-6908b9296396 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.519139] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1059.519139] env[61867]: value = "task-1277141" [ 1059.519139] env[61867]: _type = "Task" [ 1059.519139] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.526538] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277141, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.760393] env[61867]: DEBUG nova.compute.manager [req-5a6090b7-c793-4f49-9a5a-39c66bb7be8f req-7497230c-6b32-49d5-8979-93f10947424a service nova] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Received event network-changed-b7b25937-4b43-48af-88a2-628f656bbf4d {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1059.760669] env[61867]: DEBUG nova.compute.manager [req-5a6090b7-c793-4f49-9a5a-39c66bb7be8f req-7497230c-6b32-49d5-8979-93f10947424a service nova] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Refreshing instance network info cache due to event network-changed-b7b25937-4b43-48af-88a2-628f656bbf4d. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1059.760895] env[61867]: DEBUG oslo_concurrency.lockutils [req-5a6090b7-c793-4f49-9a5a-39c66bb7be8f req-7497230c-6b32-49d5-8979-93f10947424a service nova] Acquiring lock "refresh_cache-70170813-af35-4d7f-88e6-2adee247a701" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.761058] env[61867]: DEBUG oslo_concurrency.lockutils [req-5a6090b7-c793-4f49-9a5a-39c66bb7be8f req-7497230c-6b32-49d5-8979-93f10947424a service nova] Acquired lock "refresh_cache-70170813-af35-4d7f-88e6-2adee247a701" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.761227] env[61867]: DEBUG nova.network.neutron [req-5a6090b7-c793-4f49-9a5a-39c66bb7be8f req-7497230c-6b32-49d5-8979-93f10947424a service nova] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Refreshing network info cache for port b7b25937-4b43-48af-88a2-628f656bbf4d {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1059.959326] env[61867]: DEBUG nova.network.neutron [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating instance_info_cache with network_info: [{"id": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "address": "fa:16:3e:b9:62:3b", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1e84fbf-9f", "ovs_interfaceid": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.029308] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277141, 'name': CreateVM_Task, 'duration_secs': 0.307601} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.029481] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1060.030162] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.030331] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.031086] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1060.031086] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a326147-d288-4c41-ae73-3996ccb1ab24 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.035655] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1060.035655] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b83cc0-6b86-974b-3d53-afe41d393143" [ 1060.035655] env[61867]: _type = "Task" [ 1060.035655] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.043425] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b83cc0-6b86-974b-3d53-afe41d393143, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.441514] env[61867]: DEBUG nova.network.neutron [req-5a6090b7-c793-4f49-9a5a-39c66bb7be8f req-7497230c-6b32-49d5-8979-93f10947424a service nova] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Updated VIF entry in instance network info cache for port b7b25937-4b43-48af-88a2-628f656bbf4d. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1060.441883] env[61867]: DEBUG nova.network.neutron [req-5a6090b7-c793-4f49-9a5a-39c66bb7be8f req-7497230c-6b32-49d5-8979-93f10947424a service nova] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Updating instance_info_cache with network_info: [{"id": "b7b25937-4b43-48af-88a2-628f656bbf4d", "address": "fa:16:3e:df:51:fa", "network": {"id": "116eb6c3-714d-4c39-a1dc-3d18b6b8d97e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-751125907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "cedf44d3b54a42798bb06ad6aae012e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7b25937-4b", "ovs_interfaceid": "b7b25937-4b43-48af-88a2-628f656bbf4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.461096] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.461973] env[61867]: DEBUG oslo_concurrency.lockutils [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.545707] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b83cc0-6b86-974b-3d53-afe41d393143, 'name': SearchDatastore_Task, 'duration_secs': 0.009728} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.545998] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.546245] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1060.546475] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.546626] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.546805] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1060.547079] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a497df8e-268a-432c-8905-61516f55297e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.554933] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1060.555126] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1060.555820] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0c3b76a-dfb8-4d5e-9231-229471b836e9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.562166] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1060.562166] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f4dd42-b045-40e3-fef0-4d81a1a9629b" [ 1060.562166] env[61867]: _type = "Task" [ 1060.562166] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.568878] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f4dd42-b045-40e3-fef0-4d81a1a9629b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.945262] env[61867]: DEBUG oslo_concurrency.lockutils [req-5a6090b7-c793-4f49-9a5a-39c66bb7be8f req-7497230c-6b32-49d5-8979-93f10947424a service nova] Releasing lock "refresh_cache-70170813-af35-4d7f-88e6-2adee247a701" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.972328] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc90415f-5569-4b07-9873-146eb21bd372 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.979693] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9fbd6ab-6c05-4b8a-bf21-50ead3bc1341 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.073228] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f4dd42-b045-40e3-fef0-4d81a1a9629b, 'name': SearchDatastore_Task, 'duration_secs': 0.008082} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.074136] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc22dd8a-a8c8-47f8-8bb5-d7c6c176e5c4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.079069] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1061.079069] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5209e876-92b1-04b1-70d6-6d2eff0c4a81" [ 1061.079069] env[61867]: _type = "Task" [ 1061.079069] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.086512] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5209e876-92b1-04b1-70d6-6d2eff0c4a81, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.590466] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5209e876-92b1-04b1-70d6-6d2eff0c4a81, 'name': SearchDatastore_Task, 'duration_secs': 0.009463} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.590466] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.590648] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 70170813-af35-4d7f-88e6-2adee247a701/70170813-af35-4d7f-88e6-2adee247a701.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1061.590909] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-891e0bcb-1cbe-4a2d-a884-5f2ef9a9ebad {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.598149] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1061.598149] env[61867]: value = "task-1277143" [ 1061.598149] env[61867]: _type = "Task" [ 1061.598149] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.607185] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277143, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.957068] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.107743] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277143, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.109274] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e68a3a-19e0-4fcd-bba0-512d2d6e560e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.130539] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b36b507c-d07f-44dd-8f9a-dd40fc976c38 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.137379] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating instance 'df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2' progress to 83 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1062.508578] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Volume attach. Driver type: vmdk {{(pid=61867) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1062.508837] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274425', 'volume_id': 'ae2bf781-d644-4860-a51b-9a8e3602b5fd', 'name': 'volume-ae2bf781-d644-4860-a51b-9a8e3602b5fd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '748e8a71-28b6-4c91-9cbb-856359348c96', 'attached_at': '', 'detached_at': '', 'volume_id': 'ae2bf781-d644-4860-a51b-9a8e3602b5fd', 'serial': 'ae2bf781-d644-4860-a51b-9a8e3602b5fd'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1062.509738] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c5ffc0a-e960-4a2c-a4f2-863479afcba5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.528825] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c632f0a-4e01-4732-aaff-af5363f47a7a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.554101] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] volume-ae2bf781-d644-4860-a51b-9a8e3602b5fd/volume-ae2bf781-d644-4860-a51b-9a8e3602b5fd.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1062.554453] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3062ca75-c172-4693-8bcc-0651e3b6637a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.574224] env[61867]: DEBUG oslo_vmware.api [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1062.574224] env[61867]: value = "task-1277144" [ 1062.574224] env[61867]: _type = "Task" [ 1062.574224] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.582828] env[61867]: DEBUG oslo_vmware.api [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277144, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.607553] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277143, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.643753] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1062.644041] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1f3ceff5-2d1f-4a68-ad97-50816c54780c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.650747] env[61867]: DEBUG oslo_vmware.api [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1062.650747] env[61867]: value = "task-1277145" [ 1062.650747] env[61867]: _type = "Task" [ 1062.650747] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.661376] env[61867]: DEBUG oslo_vmware.api [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277145, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.083706] env[61867]: DEBUG oslo_vmware.api [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277144, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.109270] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277143, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.163815] env[61867]: DEBUG oslo_vmware.api [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277145, 'name': PowerOnVM_Task, 'duration_secs': 0.483957} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.164187] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1063.164495] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-944eba32-7917-4267-9671-0ea30cc06dcd tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating instance 'df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2' progress to 100 {{(pid=61867) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1063.585580] env[61867]: DEBUG oslo_vmware.api [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277144, 'name': ReconfigVM_Task, 'duration_secs': 0.560666} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.585877] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Reconfigured VM instance instance-00000065 to attach disk [datastore1] volume-ae2bf781-d644-4860-a51b-9a8e3602b5fd/volume-ae2bf781-d644-4860-a51b-9a8e3602b5fd.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1063.591580] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8140c826-8314-45c6-b2c3-7e5d09445bee {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.611800] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277143, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.515807} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.615388] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 70170813-af35-4d7f-88e6-2adee247a701/70170813-af35-4d7f-88e6-2adee247a701.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1063.615663] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1063.615979] env[61867]: DEBUG oslo_vmware.api [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1063.615979] env[61867]: value = "task-1277146" [ 1063.615979] env[61867]: _type = "Task" [ 1063.615979] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.616188] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad915d42-8585-43b4-82a9-23e062d1c99a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.629073] env[61867]: DEBUG oslo_vmware.api [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277146, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.630727] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1063.630727] env[61867]: value = "task-1277147" [ 1063.630727] env[61867]: _type = "Task" [ 1063.630727] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.641999] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277147, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.956780] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1063.958209] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Starting heal instance info cache {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1063.958209] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Rebuilding the list of instances to heal {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1064.127877] env[61867]: DEBUG oslo_vmware.api [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277146, 'name': ReconfigVM_Task, 'duration_secs': 0.143678} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.128256] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274425', 'volume_id': 'ae2bf781-d644-4860-a51b-9a8e3602b5fd', 'name': 'volume-ae2bf781-d644-4860-a51b-9a8e3602b5fd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '748e8a71-28b6-4c91-9cbb-856359348c96', 'attached_at': '', 'detached_at': '', 'volume_id': 'ae2bf781-d644-4860-a51b-9a8e3602b5fd', 'serial': 'ae2bf781-d644-4860-a51b-9a8e3602b5fd'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1064.140435] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277147, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062963} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.142925] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1064.142925] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cb4853d-29c2-4c85-b767-fa720c338fc6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.165549] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 70170813-af35-4d7f-88e6-2adee247a701/70170813-af35-4d7f-88e6-2adee247a701.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1064.166037] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2cebdac1-d7d3-4bf8-b4c8-232311f35ef0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.185215] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1064.185215] env[61867]: value = "task-1277148" [ 1064.185215] env[61867]: _type = "Task" [ 1064.185215] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.192710] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277148, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.461504] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Skipping network cache update for instance because it is Building. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1064.461841] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.461964] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquired lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.462114] env[61867]: DEBUG nova.network.neutron [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Forcefully refreshing network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1064.462267] env[61867]: DEBUG nova.objects.instance [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lazy-loading 'info_cache' on Instance uuid df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1064.700502] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277148, 'name': ReconfigVM_Task, 'duration_secs': 0.272252} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.700805] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 70170813-af35-4d7f-88e6-2adee247a701/70170813-af35-4d7f-88e6-2adee247a701.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.701339] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c88eeef3-6e49-487d-b53a-f7a792240f1b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.708635] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1064.708635] env[61867]: value = "task-1277149" [ 1064.708635] env[61867]: _type = "Task" [ 1064.708635] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.723048] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277149, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.177116] env[61867]: DEBUG nova.objects.instance [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lazy-loading 'flavor' on Instance uuid 748e8a71-28b6-4c91-9cbb-856359348c96 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.216351] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "6fa6c260-9c20-4808-9dd7-946a27651681" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.216525] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "6fa6c260-9c20-4808-9dd7-946a27651681" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.222128] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277149, 'name': Rename_Task, 'duration_secs': 0.156051} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.222537] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1065.222975] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f4f33602-6b53-4bb9-8ddf-2101bb78d96c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.229730] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1065.229730] env[61867]: value = "task-1277150" [ 1065.229730] env[61867]: _type = "Task" [ 1065.229730] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.239039] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277150, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.359762] env[61867]: DEBUG nova.network.neutron [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Port f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2 binding to destination host cpu-1 is already ACTIVE {{(pid=61867) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1065.359990] env[61867]: DEBUG oslo_concurrency.lockutils [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.682657] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c596a9c1-a81f-4c08-9620-99fe641e3d2d tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "748e8a71-28b6-4c91-9cbb-856359348c96" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.768s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.724745] env[61867]: DEBUG nova.compute.manager [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1065.744198] env[61867]: DEBUG oslo_vmware.api [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277150, 'name': PowerOnVM_Task, 'duration_secs': 0.424869} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.744497] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1065.744707] env[61867]: INFO nova.compute.manager [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Took 8.15 seconds to spawn the instance on the hypervisor. [ 1065.744892] env[61867]: DEBUG nova.compute.manager [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.745718] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f05d73a6-9fbe-4dea-a926-143409109284 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.200214] env[61867]: DEBUG nova.network.neutron [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating instance_info_cache with network_info: [{"id": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "address": "fa:16:3e:b9:62:3b", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1e84fbf-9f", "ovs_interfaceid": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.250881] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.251163] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.252676] env[61867]: INFO nova.compute.claims [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1066.264345] env[61867]: INFO nova.compute.manager [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Took 12.94 seconds to build instance. [ 1066.466420] env[61867]: DEBUG oslo_concurrency.lockutils [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "748e8a71-28b6-4c91-9cbb-856359348c96" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.466876] env[61867]: DEBUG oslo_concurrency.lockutils [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "748e8a71-28b6-4c91-9cbb-856359348c96" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.703168] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Releasing lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.703486] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updated the network info_cache for instance {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1066.703810] env[61867]: DEBUG oslo_concurrency.lockutils [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.704040] env[61867]: DEBUG nova.network.neutron [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1066.705302] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.705734] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.705909] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61867) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1066.706085] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.765808] env[61867]: DEBUG oslo_concurrency.lockutils [None req-d67542ee-c6e6-4e7c-8377-e3fb20e4e4d5 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "70170813-af35-4d7f-88e6-2adee247a701" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.446s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.826353] env[61867]: INFO nova.compute.manager [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Rescuing [ 1066.826597] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "refresh_cache-70170813-af35-4d7f-88e6-2adee247a701" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.826752] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquired lock "refresh_cache-70170813-af35-4d7f-88e6-2adee247a701" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.826919] env[61867]: DEBUG nova.network.neutron [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1066.970509] env[61867]: INFO nova.compute.manager [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Detaching volume ae2bf781-d644-4860-a51b-9a8e3602b5fd [ 1067.005593] env[61867]: INFO nova.virt.block_device [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Attempting to driver detach volume ae2bf781-d644-4860-a51b-9a8e3602b5fd from mountpoint /dev/sdb [ 1067.005842] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Volume detach. Driver type: vmdk {{(pid=61867) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1067.006047] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274425', 'volume_id': 'ae2bf781-d644-4860-a51b-9a8e3602b5fd', 'name': 'volume-ae2bf781-d644-4860-a51b-9a8e3602b5fd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '748e8a71-28b6-4c91-9cbb-856359348c96', 'attached_at': '', 'detached_at': '', 'volume_id': 'ae2bf781-d644-4860-a51b-9a8e3602b5fd', 'serial': 'ae2bf781-d644-4860-a51b-9a8e3602b5fd'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1067.006920] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ca19df-8903-4af9-aa85-953d4a6358ee {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.028044] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf815cd-691f-4089-aece-18e5224957b6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.034870] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa085383-6fb5-4940-9fca-eafc1be82e4a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.054687] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7438bc62-7535-4def-b206-a63a5eb94dd0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.068738] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] The volume has not been displaced from its original location: [datastore1] volume-ae2bf781-d644-4860-a51b-9a8e3602b5fd/volume-ae2bf781-d644-4860-a51b-9a8e3602b5fd.vmdk. No consolidation needed. {{(pid=61867) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1067.073905] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Reconfiguring VM instance instance-00000065 to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1067.074211] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-730c230a-de55-46cb-8fd7-126ae6568999 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.092132] env[61867]: DEBUG oslo_vmware.api [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1067.092132] env[61867]: value = "task-1277151" [ 1067.092132] env[61867]: _type = "Task" [ 1067.092132] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.100332] env[61867]: DEBUG oslo_vmware.api [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277151, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.210091] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.345919] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18013dda-55dc-497c-a117-a76118048a70 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.353165] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c2db0c-c3da-4cfd-9e13-c828afb700d0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.387310] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5d408d-d02a-4627-a62f-93796db9400b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.394296] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f622dec-58d1-4bc9-b5be-f495ee0733b1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.407975] env[61867]: DEBUG nova.compute.provider_tree [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1067.502784] env[61867]: DEBUG nova.network.neutron [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating instance_info_cache with network_info: [{"id": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "address": "fa:16:3e:b9:62:3b", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1e84fbf-9f", "ovs_interfaceid": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.572376] env[61867]: DEBUG nova.network.neutron [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Updating instance_info_cache with network_info: [{"id": "b7b25937-4b43-48af-88a2-628f656bbf4d", "address": "fa:16:3e:df:51:fa", "network": {"id": "116eb6c3-714d-4c39-a1dc-3d18b6b8d97e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-751125907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "cedf44d3b54a42798bb06ad6aae012e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7b25937-4b", "ovs_interfaceid": "b7b25937-4b43-48af-88a2-628f656bbf4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.601282] env[61867]: DEBUG oslo_vmware.api [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277151, 'name': ReconfigVM_Task, 'duration_secs': 0.225647} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.601582] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Reconfigured VM instance instance-00000065 to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1067.606170] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-122ce63d-4177-43b0-9c80-d422044d0a30 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.620557] env[61867]: DEBUG oslo_vmware.api [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1067.620557] env[61867]: value = "task-1277152" [ 1067.620557] env[61867]: _type = "Task" [ 1067.620557] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.627730] env[61867]: DEBUG oslo_vmware.api [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277152, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.911026] env[61867]: DEBUG nova.scheduler.client.report [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1068.005400] env[61867]: DEBUG oslo_concurrency.lockutils [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.074791] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Releasing lock "refresh_cache-70170813-af35-4d7f-88e6-2adee247a701" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.130997] env[61867]: DEBUG oslo_vmware.api [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277152, 'name': ReconfigVM_Task, 'duration_secs': 0.139046} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.131335] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274425', 'volume_id': 'ae2bf781-d644-4860-a51b-9a8e3602b5fd', 'name': 'volume-ae2bf781-d644-4860-a51b-9a8e3602b5fd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '748e8a71-28b6-4c91-9cbb-856359348c96', 'attached_at': '', 'detached_at': '', 'volume_id': 'ae2bf781-d644-4860-a51b-9a8e3602b5fd', 'serial': 'ae2bf781-d644-4860-a51b-9a8e3602b5fd'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1068.416370] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.165s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.416856] env[61867]: DEBUG nova.compute.manager [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1068.419453] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.209s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.419563] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.419717] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61867) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1068.420690] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6bcf52b-a7d4-45f4-90f9-166bdae400f1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.428916] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eddaeb9b-780b-4d78-851c-d9adb28d727b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.447790] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b43ab4-4275-47b5-8a3b-8821d3bbcafd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.458380] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ceb4730-03da-4c40-9c20-5d1c92436d8f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.495484] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180909MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61867) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1068.495637] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.495824] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.509217] env[61867]: DEBUG nova.compute.manager [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61867) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1068.604471] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1068.606031] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1b45c169-1abf-477d-96b3-b9611297f853 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.611865] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1068.611865] env[61867]: value = "task-1277153" [ 1068.611865] env[61867]: _type = "Task" [ 1068.611865] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.620082] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277153, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.670689] env[61867]: DEBUG nova.objects.instance [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lazy-loading 'flavor' on Instance uuid 748e8a71-28b6-4c91-9cbb-856359348c96 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.922444] env[61867]: DEBUG nova.compute.utils [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1068.923911] env[61867]: DEBUG nova.compute.manager [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1068.924459] env[61867]: DEBUG nova.network.neutron [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1068.960651] env[61867]: DEBUG nova.policy [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bcdb4a387cdb45899eded851c1bc842a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f245b07249fa4e37ac97cf66dbf6678f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 1069.121289] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277153, 'name': PowerOffVM_Task, 'duration_secs': 0.199534} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.121586] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1069.122351] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56455fbb-8648-4261-8098-c8b8f9dbcd5f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.140103] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781390ac-d3ec-49d5-94e0-4d2f7143fefd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.166253] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1069.166512] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e97eefd-168f-4a07-8e91-8ddf1e134ac4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.172442] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1069.172442] env[61867]: value = "task-1277154" [ 1069.172442] env[61867]: _type = "Task" [ 1069.172442] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.182291] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] VM already powered off {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1069.182461] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1069.182701] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.182851] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.183036] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1069.183297] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a7ae4bea-d229-47f8-9242-d6d2f06e5eba {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.190877] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1069.191425] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1069.191855] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64b02062-03ca-4e07-9cf6-3229735033a6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.196756] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1069.196756] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52164872-2bd3-e1fb-a9b4-2e90847a7235" [ 1069.196756] env[61867]: _type = "Task" [ 1069.196756] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.203791] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52164872-2bd3-e1fb-a9b4-2e90847a7235, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.365497] env[61867]: DEBUG nova.network.neutron [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Successfully created port: 5bc71c96-38d5-4cd8-92d4-46e8713e8389 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1069.427712] env[61867]: DEBUG nova.compute.manager [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1069.503298] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Applying migration context for instance df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2 as it has an incoming, in-progress migration b428c0ea-e305-41b1-bae7-2d1c68bd7059. Migration status is reverting {{(pid=61867) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1069.504216] env[61867]: INFO nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating resource usage from migration b428c0ea-e305-41b1-bae7-2d1c68bd7059 [ 1069.593789] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 748e8a71-28b6-4c91-9cbb-856359348c96 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.593897] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance a9724422-abd1-4771-acb2-2c0ac45f5c16 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.594107] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Migration b428c0ea-e305-41b1-bae7-2d1c68bd7059 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1069.594243] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.594432] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 70170813-af35-4d7f-88e6-2adee247a701 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.594618] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 6fa6c260-9c20-4808-9dd7-946a27651681 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.594877] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1069.595029] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1728MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1069.597441] env[61867]: DEBUG oslo_concurrency.lockutils [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.667398] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ab361f-2c79-45f8-8860-dac7405505d6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.674572] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-601202df-cece-4477-9e37-2fad84d4953a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.679325] env[61867]: DEBUG oslo_concurrency.lockutils [None req-49ba13f2-72da-4c9e-a307-d9108a9b58cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "748e8a71-28b6-4c91-9cbb-856359348c96" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.213s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.709561] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36a785f7-e0fb-4fd2-9436-9321a7c79fc5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.717670] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52164872-2bd3-e1fb-a9b4-2e90847a7235, 'name': SearchDatastore_Task, 'duration_secs': 0.008398} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.720082] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-539d61de-8df7-407d-8912-78782904767b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.722981] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9792f014-d3d8-4e30-a502-17ffe77dd9bf {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.737138] env[61867]: DEBUG nova.compute.provider_tree [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1069.739975] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1069.739975] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d8812a-bcb4-8e67-0cad-50d4d1eec3af" [ 1069.739975] env[61867]: _type = "Task" [ 1069.739975] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.747429] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d8812a-bcb4-8e67-0cad-50d4d1eec3af, 'name': SearchDatastore_Task, 'duration_secs': 0.009563} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.748170] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.748428] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 70170813-af35-4d7f-88e6-2adee247a701/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk. {{(pid=61867) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1069.748670] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-131f6d5f-102b-4a64-be5a-1dee89ec01e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.755519] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1069.755519] env[61867]: value = "task-1277155" [ 1069.755519] env[61867]: _type = "Task" [ 1069.755519] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.762847] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277155, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.241085] env[61867]: DEBUG nova.scheduler.client.report [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1070.264998] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277155, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467073} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.265269] env[61867]: INFO nova.virt.vmwareapi.ds_util [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 70170813-af35-4d7f-88e6-2adee247a701/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk. [ 1070.266042] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af372a27-7d5f-48eb-84ed-72b7d3b7e9c3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.290418] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 70170813-af35-4d7f-88e6-2adee247a701/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1070.290676] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e4d871e-2174-4342-8356-48daf99f6240 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.310352] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1070.310352] env[61867]: value = "task-1277156" [ 1070.310352] env[61867]: _type = "Task" [ 1070.310352] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.318076] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277156, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.437712] env[61867]: DEBUG nova.compute.manager [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1070.461912] env[61867]: DEBUG nova.virt.hardware [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1070.462195] env[61867]: DEBUG nova.virt.hardware [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1070.462360] env[61867]: DEBUG nova.virt.hardware [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1070.462549] env[61867]: DEBUG nova.virt.hardware [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1070.462699] env[61867]: DEBUG nova.virt.hardware [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1070.462848] env[61867]: DEBUG nova.virt.hardware [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1070.463076] env[61867]: DEBUG nova.virt.hardware [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1070.463251] env[61867]: DEBUG nova.virt.hardware [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1070.463447] env[61867]: DEBUG nova.virt.hardware [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1070.463621] env[61867]: DEBUG nova.virt.hardware [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1070.463799] env[61867]: DEBUG nova.virt.hardware [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1070.464701] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bfc03b3-fa7e-48d9-bab4-bbbfeb69d7bd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.472800] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cccdb8f7-d1ef-49f3-a70d-d490811fca02 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.728847] env[61867]: DEBUG oslo_concurrency.lockutils [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "748e8a71-28b6-4c91-9cbb-856359348c96" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.729178] env[61867]: DEBUG oslo_concurrency.lockutils [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "748e8a71-28b6-4c91-9cbb-856359348c96" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.729367] env[61867]: DEBUG oslo_concurrency.lockutils [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "748e8a71-28b6-4c91-9cbb-856359348c96-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.729563] env[61867]: DEBUG oslo_concurrency.lockutils [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "748e8a71-28b6-4c91-9cbb-856359348c96-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.729739] env[61867]: DEBUG oslo_concurrency.lockutils [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "748e8a71-28b6-4c91-9cbb-856359348c96-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.731985] env[61867]: INFO nova.compute.manager [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Terminating instance [ 1070.733961] env[61867]: DEBUG nova.compute.manager [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1070.734193] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1070.735154] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e376f4-10b4-4035-817d-1e11551c12ea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.739848] env[61867]: DEBUG nova.compute.manager [req-f9c0d1c4-c758-490d-b7ca-a5fe8dba36a9 req-03cba4eb-dacf-4071-a0cc-a74ccc1f8e96 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Received event network-vif-plugged-5bc71c96-38d5-4cd8-92d4-46e8713e8389 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1070.740071] env[61867]: DEBUG oslo_concurrency.lockutils [req-f9c0d1c4-c758-490d-b7ca-a5fe8dba36a9 req-03cba4eb-dacf-4071-a0cc-a74ccc1f8e96 service nova] Acquiring lock "6fa6c260-9c20-4808-9dd7-946a27651681-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.740278] env[61867]: DEBUG oslo_concurrency.lockutils [req-f9c0d1c4-c758-490d-b7ca-a5fe8dba36a9 req-03cba4eb-dacf-4071-a0cc-a74ccc1f8e96 service nova] Lock "6fa6c260-9c20-4808-9dd7-946a27651681-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.740452] env[61867]: DEBUG oslo_concurrency.lockutils [req-f9c0d1c4-c758-490d-b7ca-a5fe8dba36a9 req-03cba4eb-dacf-4071-a0cc-a74ccc1f8e96 service nova] Lock "6fa6c260-9c20-4808-9dd7-946a27651681-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.740624] env[61867]: DEBUG nova.compute.manager [req-f9c0d1c4-c758-490d-b7ca-a5fe8dba36a9 req-03cba4eb-dacf-4071-a0cc-a74ccc1f8e96 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] No waiting events found dispatching network-vif-plugged-5bc71c96-38d5-4cd8-92d4-46e8713e8389 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1070.740796] env[61867]: WARNING nova.compute.manager [req-f9c0d1c4-c758-490d-b7ca-a5fe8dba36a9 req-03cba4eb-dacf-4071-a0cc-a74ccc1f8e96 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Received unexpected event network-vif-plugged-5bc71c96-38d5-4cd8-92d4-46e8713e8389 for instance with vm_state building and task_state spawning. [ 1070.746420] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61867) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1070.746596] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.251s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.746842] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1070.747123] env[61867]: DEBUG oslo_concurrency.lockutils [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 1.150s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.748261] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1070.748407] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Cleaning up deleted instances {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1070.749434] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3806b901-9693-4edb-9e43-fea2ea489c80 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.755696] env[61867]: DEBUG oslo_vmware.api [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1070.755696] env[61867]: value = "task-1277157" [ 1070.755696] env[61867]: _type = "Task" [ 1070.755696] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.764752] env[61867]: DEBUG oslo_vmware.api [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277157, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.821287] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277156, 'name': ReconfigVM_Task, 'duration_secs': 0.273915} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.821600] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 70170813-af35-4d7f-88e6-2adee247a701/4ca02567-c128-482f-b204-6ac166ac3160-rescue.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1070.822616] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f55a480-e806-4de0-a7c4-f82f9430fb87 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.849202] env[61867]: DEBUG nova.network.neutron [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Successfully updated port: 5bc71c96-38d5-4cd8-92d4-46e8713e8389 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1070.850436] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da3aaa1c-bc4e-4c85-bdd0-edc8d82a5d8e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.869129] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1070.869129] env[61867]: value = "task-1277158" [ 1070.869129] env[61867]: _type = "Task" [ 1070.869129] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.878658] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277158, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.252347] env[61867]: DEBUG nova.objects.instance [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lazy-loading 'migration_context' on Instance uuid df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.270108] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] There are 52 instances to clean {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1071.270383] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 64947dbe-062c-4041-9ba9-d055279ded1e] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.284662] env[61867]: DEBUG oslo_vmware.api [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277157, 'name': PowerOffVM_Task, 'duration_secs': 0.184374} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.284982] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1071.285165] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1071.285647] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-96d1a9eb-bbb2-4abe-b1af-1b3f3d5c7186 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.347406] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1071.347638] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1071.347818] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Deleting the datastore file [datastore2] 748e8a71-28b6-4c91-9cbb-856359348c96 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1071.348093] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3046d73c-5cb5-44ba-8bad-d4e71e2eaa6a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.354056] env[61867]: DEBUG oslo_vmware.api [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1071.354056] env[61867]: value = "task-1277160" [ 1071.354056] env[61867]: _type = "Task" [ 1071.354056] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.361694] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.361836] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquired lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.361926] env[61867]: DEBUG nova.network.neutron [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1071.363018] env[61867]: DEBUG oslo_vmware.api [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277160, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.379480] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277158, 'name': ReconfigVM_Task, 'duration_secs': 0.161733} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.379949] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1071.380202] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4cf92073-7e8c-40fd-a64b-e09d69e4ac26 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.386914] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1071.386914] env[61867]: value = "task-1277161" [ 1071.386914] env[61867]: _type = "Task" [ 1071.386914] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.396161] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277161, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.776801] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 986ada58-eb26-4289-bc0e-338c6a399645] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.861322] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ef844e-301c-4c31-bbe6-a2306e0661a8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.869315] env[61867]: DEBUG oslo_vmware.api [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277160, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.201568} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.869952] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1071.870165] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1071.870354] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1071.870530] env[61867]: INFO nova.compute.manager [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1071.870767] env[61867]: DEBUG oslo.service.loopingcall [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1071.872045] env[61867]: DEBUG nova.compute.manager [-] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1071.872045] env[61867]: DEBUG nova.network.neutron [-] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1071.875618] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c85bc6-6be1-4fa6-b1a5-c9a9e01d8984 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.908054] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e7acb1e-a81e-471c-9cdb-08044de99bd3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.916127] env[61867]: DEBUG oslo_vmware.api [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277161, 'name': PowerOnVM_Task, 'duration_secs': 0.406925} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.916844] env[61867]: DEBUG nova.network.neutron [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1071.922183] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1071.926916] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0994ca2-e9cc-4d87-aa97-92f3e942fba2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.932785] env[61867]: DEBUG nova.compute.manager [None req-0a2b6aa6-482a-4fe6-9a9c-bd75cb41badf tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1071.933717] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22f22b60-e80b-43a1-99ed-bf52b3ab85d7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.947133] env[61867]: DEBUG nova.compute.provider_tree [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1072.147418] env[61867]: DEBUG nova.network.neutron [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Updating instance_info_cache with network_info: [{"id": "5bc71c96-38d5-4cd8-92d4-46e8713e8389", "address": "fa:16:3e:55:cd:c4", "network": {"id": "8db0c704-eb45-4fca-8624-d21877ab5ab3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-74001547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f245b07249fa4e37ac97cf66dbf6678f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bc71c96-38", "ovs_interfaceid": "5bc71c96-38d5-4cd8-92d4-46e8713e8389", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.280731] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 29d961df-5db5-46db-ace9-651913e0250a] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.456043] env[61867]: DEBUG nova.scheduler.client.report [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1072.650538] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Releasing lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.651191] env[61867]: DEBUG nova.compute.manager [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Instance network_info: |[{"id": "5bc71c96-38d5-4cd8-92d4-46e8713e8389", "address": "fa:16:3e:55:cd:c4", "network": {"id": "8db0c704-eb45-4fca-8624-d21877ab5ab3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-74001547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f245b07249fa4e37ac97cf66dbf6678f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bc71c96-38", "ovs_interfaceid": "5bc71c96-38d5-4cd8-92d4-46e8713e8389", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1072.651333] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:cd:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91c1da19-ab68-4127-bacd-accbaff19651', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5bc71c96-38d5-4cd8-92d4-46e8713e8389', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1072.659026] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Creating folder: Project (f245b07249fa4e37ac97cf66dbf6678f). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1072.659273] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b66e9955-9ea6-477c-b7c1-0de1a8c44161 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.671022] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Created folder: Project (f245b07249fa4e37ac97cf66dbf6678f) in parent group-v274258. [ 1072.671226] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Creating folder: Instances. Parent ref: group-v274427. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1072.671469] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a26b2584-80e8-4efd-9354-20adb7a7ba51 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.681142] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Created folder: Instances in parent group-v274427. [ 1072.681390] env[61867]: DEBUG oslo.service.loopingcall [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1072.681584] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1072.681784] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a11b10f0-58fa-433b-b13f-f2b5c3346491 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.699886] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1072.699886] env[61867]: value = "task-1277164" [ 1072.699886] env[61867]: _type = "Task" [ 1072.699886] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.707092] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277164, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.770960] env[61867]: DEBUG nova.compute.manager [req-a1c4f7ea-1a41-45b9-b468-902ce22aa127 req-49d42dc7-8bbc-4aa1-99e0-bdc4cc371853 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Received event network-changed-5bc71c96-38d5-4cd8-92d4-46e8713e8389 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1072.771222] env[61867]: DEBUG nova.compute.manager [req-a1c4f7ea-1a41-45b9-b468-902ce22aa127 req-49d42dc7-8bbc-4aa1-99e0-bdc4cc371853 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Refreshing instance network info cache due to event network-changed-5bc71c96-38d5-4cd8-92d4-46e8713e8389. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1072.771551] env[61867]: DEBUG oslo_concurrency.lockutils [req-a1c4f7ea-1a41-45b9-b468-902ce22aa127 req-49d42dc7-8bbc-4aa1-99e0-bdc4cc371853 service nova] Acquiring lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.771710] env[61867]: DEBUG oslo_concurrency.lockutils [req-a1c4f7ea-1a41-45b9-b468-902ce22aa127 req-49d42dc7-8bbc-4aa1-99e0-bdc4cc371853 service nova] Acquired lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.771878] env[61867]: DEBUG nova.network.neutron [req-a1c4f7ea-1a41-45b9-b468-902ce22aa127 req-49d42dc7-8bbc-4aa1-99e0-bdc4cc371853 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Refreshing network info cache for port 5bc71c96-38d5-4cd8-92d4-46e8713e8389 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1072.784524] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 4006cabd-cb6e-45b6-b87e-9b4b1550f4fe] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.935164] env[61867]: DEBUG nova.network.neutron [-] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.209767] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277164, 'name': CreateVM_Task, 'duration_secs': 0.337382} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.209967] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1073.210730] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.210902] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.211234] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1073.212027] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7a5a414-189c-43de-9063-034676db6cca {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.216261] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1073.216261] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527c1b9b-c73a-28f3-d455-ad219ca2a5bc" [ 1073.216261] env[61867]: _type = "Task" [ 1073.216261] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.224309] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527c1b9b-c73a-28f3-d455-ad219ca2a5bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.287016] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 976299f8-c6cb-4106-90f2-0a4c85625d8c] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1073.438018] env[61867]: INFO nova.compute.manager [-] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Took 1.57 seconds to deallocate network for instance. [ 1073.461657] env[61867]: DEBUG nova.network.neutron [req-a1c4f7ea-1a41-45b9-b468-902ce22aa127 req-49d42dc7-8bbc-4aa1-99e0-bdc4cc371853 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Updated VIF entry in instance network info cache for port 5bc71c96-38d5-4cd8-92d4-46e8713e8389. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1073.462019] env[61867]: DEBUG nova.network.neutron [req-a1c4f7ea-1a41-45b9-b468-902ce22aa127 req-49d42dc7-8bbc-4aa1-99e0-bdc4cc371853 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Updating instance_info_cache with network_info: [{"id": "5bc71c96-38d5-4cd8-92d4-46e8713e8389", "address": "fa:16:3e:55:cd:c4", "network": {"id": "8db0c704-eb45-4fca-8624-d21877ab5ab3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-74001547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f245b07249fa4e37ac97cf66dbf6678f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bc71c96-38", "ovs_interfaceid": "5bc71c96-38d5-4cd8-92d4-46e8713e8389", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.465662] env[61867]: DEBUG oslo_concurrency.lockutils [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.718s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.657430] env[61867]: INFO nova.compute.manager [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Unrescuing [ 1073.657695] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "refresh_cache-70170813-af35-4d7f-88e6-2adee247a701" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.657854] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquired lock "refresh_cache-70170813-af35-4d7f-88e6-2adee247a701" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.658032] env[61867]: DEBUG nova.network.neutron [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1073.726636] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527c1b9b-c73a-28f3-d455-ad219ca2a5bc, 'name': SearchDatastore_Task, 'duration_secs': 0.009496} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.726937] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.727214] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1073.727455] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.727610] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.727793] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1073.728063] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-74010455-d83b-4e95-b200-09fc4972c061 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.736275] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1073.736456] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1073.737165] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37d0a331-1cfd-428e-9c86-fbb4750a9995 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.741808] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1073.741808] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b60f48-5c62-db87-bb6f-e9933938aac1" [ 1073.741808] env[61867]: _type = "Task" [ 1073.741808] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.749017] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b60f48-5c62-db87-bb6f-e9933938aac1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.790807] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: d19e7706-e6ac-47dd-8ec3-c30bd315559d] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1073.944147] env[61867]: DEBUG oslo_concurrency.lockutils [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.944393] env[61867]: DEBUG oslo_concurrency.lockutils [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.944633] env[61867]: DEBUG nova.objects.instance [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lazy-loading 'resources' on Instance uuid 748e8a71-28b6-4c91-9cbb-856359348c96 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1073.963996] env[61867]: DEBUG oslo_concurrency.lockutils [req-a1c4f7ea-1a41-45b9-b468-902ce22aa127 req-49d42dc7-8bbc-4aa1-99e0-bdc4cc371853 service nova] Releasing lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.964261] env[61867]: DEBUG nova.compute.manager [req-a1c4f7ea-1a41-45b9-b468-902ce22aa127 req-49d42dc7-8bbc-4aa1-99e0-bdc4cc371853 service nova] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Received event network-vif-deleted-eb146e5d-990b-477a-928e-ba9014b0fb40 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1073.964483] env[61867]: INFO nova.compute.manager [req-a1c4f7ea-1a41-45b9-b468-902ce22aa127 req-49d42dc7-8bbc-4aa1-99e0-bdc4cc371853 service nova] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Neutron deleted interface eb146e5d-990b-477a-928e-ba9014b0fb40; detaching it from the instance and deleting it from the info cache [ 1073.964669] env[61867]: DEBUG nova.network.neutron [req-a1c4f7ea-1a41-45b9-b468-902ce22aa127 req-49d42dc7-8bbc-4aa1-99e0-bdc4cc371853 service nova] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.252375] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52b60f48-5c62-db87-bb6f-e9933938aac1, 'name': SearchDatastore_Task, 'duration_secs': 0.008987} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.255128] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59540ada-3aea-4a0c-933e-4a0112a75443 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.260660] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1074.260660] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52a6a9d8-499a-e5f0-fb84-51a25f690ca8" [ 1074.260660] env[61867]: _type = "Task" [ 1074.260660] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.269462] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52a6a9d8-499a-e5f0-fb84-51a25f690ca8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.294637] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: f426dd3e-2806-4d70-8784-75c611fd03e4] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1074.335086] env[61867]: DEBUG nova.network.neutron [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Updating instance_info_cache with network_info: [{"id": "b7b25937-4b43-48af-88a2-628f656bbf4d", "address": "fa:16:3e:df:51:fa", "network": {"id": "116eb6c3-714d-4c39-a1dc-3d18b6b8d97e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-751125907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "cedf44d3b54a42798bb06ad6aae012e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7b25937-4b", "ovs_interfaceid": "b7b25937-4b43-48af-88a2-628f656bbf4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.467380] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8947ca43-9c98-43fc-9e37-31286a809d41 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.476149] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197ad4c5-e5ba-405c-b127-a1e7396f040d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.510332] env[61867]: DEBUG nova.compute.manager [req-a1c4f7ea-1a41-45b9-b468-902ce22aa127 req-49d42dc7-8bbc-4aa1-99e0-bdc4cc371853 service nova] [instance: 748e8a71-28b6-4c91-9cbb-856359348c96] Detach interface failed, port_id=eb146e5d-990b-477a-928e-ba9014b0fb40, reason: Instance 748e8a71-28b6-4c91-9cbb-856359348c96 could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1074.551646] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-791ca93e-c11f-4bdb-8d77-f11b596fcf93 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.558669] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b1e47a7-1a68-454d-ba96-32a4dbb171e6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.587101] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a563f6e0-545e-4d76-a04a-5d8089b5c47c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.593895] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3214ee55-e86d-403d-9eb2-dd557e7f29ee {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.607694] env[61867]: DEBUG nova.compute.provider_tree [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1074.771043] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52a6a9d8-499a-e5f0-fb84-51a25f690ca8, 'name': SearchDatastore_Task, 'duration_secs': 0.00897} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.771043] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.771234] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 6fa6c260-9c20-4808-9dd7-946a27651681/6fa6c260-9c20-4808-9dd7-946a27651681.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1074.771341] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3bfafa06-e766-4b50-a7bc-9aa7986cbca7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.777886] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1074.777886] env[61867]: value = "task-1277165" [ 1074.777886] env[61867]: _type = "Task" [ 1074.777886] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.785275] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277165, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.797833] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 0ae83268-ef50-42e6-954d-5e67c4afc20e] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1074.838043] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Releasing lock "refresh_cache-70170813-af35-4d7f-88e6-2adee247a701" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.838043] env[61867]: DEBUG nova.objects.instance [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lazy-loading 'flavor' on Instance uuid 70170813-af35-4d7f-88e6-2adee247a701 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1075.010574] env[61867]: INFO nova.compute.manager [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Swapping old allocation on dict_keys(['25720271-a549-4916-abe3-e5ed9b765889']) held by migration b428c0ea-e305-41b1-bae7-2d1c68bd7059 for instance [ 1075.036201] env[61867]: DEBUG nova.scheduler.client.report [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Overwriting current allocation {'allocations': {'25720271-a549-4916-abe3-e5ed9b765889': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 138}}, 'project_id': '250e5d2fa5cb4a8a8bd0a6fdf9e50a3d', 'user_id': '1fc8ff85bced44c3a9de8b6c71d02bd5', 'consumer_generation': 1} on consumer df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2 {{(pid=61867) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1075.110368] env[61867]: DEBUG nova.scheduler.client.report [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1075.130047] env[61867]: DEBUG oslo_concurrency.lockutils [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.130336] env[61867]: DEBUG oslo_concurrency.lockutils [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.130580] env[61867]: DEBUG nova.network.neutron [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1075.290112] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277165, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508085} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.290112] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 6fa6c260-9c20-4808-9dd7-946a27651681/6fa6c260-9c20-4808-9dd7-946a27651681.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1075.290112] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1075.290112] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e115e869-0726-4c77-b662-544d3459d903 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.295954] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1075.295954] env[61867]: value = "task-1277166" [ 1075.295954] env[61867]: _type = "Task" [ 1075.295954] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.300407] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 6e2136e7-c7e0-4a98-9899-f79c10f0e703] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.305549] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277166, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.343047] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f1dd1c2-f6aa-422f-8db9-9581dec12c87 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.364562] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1075.364860] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4696f4fd-4360-497e-80d8-f25d53cbfa55 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.371078] env[61867]: DEBUG oslo_vmware.api [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1075.371078] env[61867]: value = "task-1277167" [ 1075.371078] env[61867]: _type = "Task" [ 1075.371078] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.378548] env[61867]: DEBUG oslo_vmware.api [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277167, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.617054] env[61867]: DEBUG oslo_concurrency.lockutils [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.672s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.636352] env[61867]: INFO nova.scheduler.client.report [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Deleted allocations for instance 748e8a71-28b6-4c91-9cbb-856359348c96 [ 1075.806711] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 5cf15593-54f6-405c-8435-2e3b378983b7] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.808588] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277166, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088768} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.809035] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1075.809787] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93712797-7837-4b44-8a02-7414e82f12e2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.832546] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 6fa6c260-9c20-4808-9dd7-946a27651681/6fa6c260-9c20-4808-9dd7-946a27651681.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1075.833387] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aaedbf1f-7d06-4a59-b92d-1890c5020de0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.853314] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1075.853314] env[61867]: value = "task-1277168" [ 1075.853314] env[61867]: _type = "Task" [ 1075.853314] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.856961] env[61867]: DEBUG nova.network.neutron [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating instance_info_cache with network_info: [{"id": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "address": "fa:16:3e:b9:62:3b", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1e84fbf-9f", "ovs_interfaceid": "f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.863473] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277168, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.880198] env[61867]: DEBUG oslo_vmware.api [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277167, 'name': PowerOffVM_Task, 'duration_secs': 0.199408} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.880198] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1075.888100] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Reconfiguring VM instance instance-0000006a to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1075.888100] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd564726-2ffc-4f1d-8d3b-0c91c439dc53 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.904244] env[61867]: DEBUG oslo_vmware.api [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1075.904244] env[61867]: value = "task-1277169" [ 1075.904244] env[61867]: _type = "Task" [ 1075.904244] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.912264] env[61867]: DEBUG oslo_vmware.api [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277169, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.145078] env[61867]: DEBUG oslo_concurrency.lockutils [None req-447c6188-1c6a-4fa5-85c1-72b79165619b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "748e8a71-28b6-4c91-9cbb-856359348c96" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.416s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.311531] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 43dc6520-5256-4b5f-a273-6c9e0e407c72] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.359962] env[61867]: DEBUG oslo_concurrency.lockutils [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "refresh_cache-df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.360847] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f56673d6-3669-4034-84c9-93b8912b501f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.366552] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277168, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.370347] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6268043b-868e-472d-a166-e24535691128 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.414580] env[61867]: DEBUG oslo_vmware.api [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277169, 'name': ReconfigVM_Task, 'duration_secs': 0.208166} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.414889] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Reconfigured VM instance instance-0000006a to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1076.415110] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1076.415375] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f3b920b7-aca6-4390-918d-9e52352e6f92 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.421960] env[61867]: DEBUG oslo_vmware.api [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1076.421960] env[61867]: value = "task-1277170" [ 1076.421960] env[61867]: _type = "Task" [ 1076.421960] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.429702] env[61867]: DEBUG oslo_vmware.api [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277170, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.813591] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 4d7bfb31-d565-49e4-8c51-0122acebba2d] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.863239] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277168, 'name': ReconfigVM_Task, 'duration_secs': 0.799477} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.863540] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 6fa6c260-9c20-4808-9dd7-946a27651681/6fa6c260-9c20-4808-9dd7-946a27651681.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1076.864232] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c4f0a962-13dc-4398-9e5f-c329f714f4c3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.870887] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1076.870887] env[61867]: value = "task-1277171" [ 1076.870887] env[61867]: _type = "Task" [ 1076.870887] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.880905] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277171, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.934542] env[61867]: DEBUG oslo_vmware.api [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277170, 'name': PowerOnVM_Task, 'duration_secs': 0.426906} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.934797] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1076.935042] env[61867]: DEBUG nova.compute.manager [None req-f8b99286-acb5-4ebb-8682-2b336c1e8054 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1076.935859] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c69db6-ad99-45f6-bfef-c8ada165874d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.317416] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 3905193e-04da-439a-bf6c-16f638a692bf] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.385156] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277171, 'name': Rename_Task, 'duration_secs': 0.131049} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.385609] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1077.385981] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cdb5a173-92b7-4a84-95f2-ac37715ee6a7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.393159] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1077.393159] env[61867]: value = "task-1277172" [ 1077.393159] env[61867]: _type = "Task" [ 1077.393159] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.404433] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277172, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.453732] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1077.453964] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4ac2b1a8-3409-490c-a623-6e3aba176490 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.460551] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1077.460551] env[61867]: value = "task-1277173" [ 1077.460551] env[61867]: _type = "Task" [ 1077.460551] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.469346] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277173, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.741617] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.741822] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.822088] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 8a7f4314-0fd4-49f9-8eb6-12baa0977a53] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.902438] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277172, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.970267] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277173, 'name': PowerOffVM_Task, 'duration_secs': 0.188082} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.970506] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1077.971199] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1077.971455] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1077.971631] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1077.971817] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1077.971968] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1077.972166] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1077.972383] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1077.972543] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1077.972710] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1077.972868] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1077.973052] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1077.978296] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3f5576b-5497-4cba-8b38-614c109d6209 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.996912] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1077.996912] env[61867]: value = "task-1277174" [ 1077.996912] env[61867]: _type = "Task" [ 1077.996912] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.005541] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277174, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.244393] env[61867]: DEBUG nova.compute.manager [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1078.325499] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 38edb89a-28e4-4dd2-a8aa-35cb95a72e0f] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.402406] env[61867]: DEBUG oslo_vmware.api [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277172, 'name': PowerOnVM_Task, 'duration_secs': 0.521962} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.402696] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1078.402901] env[61867]: INFO nova.compute.manager [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Took 7.97 seconds to spawn the instance on the hypervisor. [ 1078.403115] env[61867]: DEBUG nova.compute.manager [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1078.403890] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4304a51-1213-47ca-9191-e76432960503 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.493900] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "70170813-af35-4d7f-88e6-2adee247a701" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.494065] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "70170813-af35-4d7f-88e6-2adee247a701" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.494277] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "70170813-af35-4d7f-88e6-2adee247a701-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.494476] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "70170813-af35-4d7f-88e6-2adee247a701-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.494676] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "70170813-af35-4d7f-88e6-2adee247a701-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.496682] env[61867]: INFO nova.compute.manager [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Terminating instance [ 1078.501263] env[61867]: DEBUG nova.compute.manager [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1078.501459] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1078.502236] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-670667e7-cab4-4d0b-9af7-5d6b2024ead8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.510095] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277174, 'name': ReconfigVM_Task, 'duration_secs': 0.240796} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.512142] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1078.512793] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b14651-d6ab-4b5d-94cc-4f7890e93699 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.515013] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-872da0d9-5285-460f-906b-13c54f4c2850 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.534823] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1078.535070] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1078.535239] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1078.535426] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1078.535596] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1078.535787] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1078.536012] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1078.536178] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1078.536349] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1078.536516] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1078.536694] env[61867]: DEBUG nova.virt.hardware [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1078.538645] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f063c234-ac2a-4ed5-be31-9931bd67504e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.540992] env[61867]: DEBUG oslo_vmware.api [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1078.540992] env[61867]: value = "task-1277175" [ 1078.540992] env[61867]: _type = "Task" [ 1078.540992] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.545325] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1078.545325] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d342f9-e234-6c5b-eefd-4771e5fdaa01" [ 1078.545325] env[61867]: _type = "Task" [ 1078.545325] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.551312] env[61867]: DEBUG oslo_vmware.api [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277175, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.555905] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d342f9-e234-6c5b-eefd-4771e5fdaa01, 'name': SearchDatastore_Task, 'duration_secs': 0.006116} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.561445] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Reconfiguring VM instance instance-00000064 to detach disk 2000 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1078.561699] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-280f10f6-1981-411c-b988-83c2567fa9a4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.579583] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1078.579583] env[61867]: value = "task-1277176" [ 1078.579583] env[61867]: _type = "Task" [ 1078.579583] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.587175] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277176, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.768329] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.768595] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.770322] env[61867]: INFO nova.compute.claims [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1078.828525] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 6d20f463-9198-4590-8ec2-db471c64ba7c] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.921076] env[61867]: INFO nova.compute.manager [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Took 12.69 seconds to build instance. [ 1079.052696] env[61867]: DEBUG oslo_vmware.api [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277175, 'name': PowerOffVM_Task, 'duration_secs': 0.204205} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.052996] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1079.053186] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1079.053458] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a3d1ab3-b007-42fd-a5ca-d40abb5491c4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.088993] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277176, 'name': ReconfigVM_Task, 'duration_secs': 0.198955} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.089289] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Reconfigured VM instance instance-00000064 to detach disk 2000 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1079.090048] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ada4d11-4500-4800-88a6-88f99d2890c2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.115251] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2/df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1079.116726] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c20e896-0eda-4d19-aeb4-35ca3c7537fe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.131501] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1079.131726] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1079.131907] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Deleting the datastore file [datastore1] 70170813-af35-4d7f-88e6-2adee247a701 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1079.132207] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0ff4700b-049f-422e-bfdb-88c51f9dc8da {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.138922] env[61867]: DEBUG oslo_vmware.api [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1079.138922] env[61867]: value = "task-1277179" [ 1079.138922] env[61867]: _type = "Task" [ 1079.138922] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.140162] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1079.140162] env[61867]: value = "task-1277178" [ 1079.140162] env[61867]: _type = "Task" [ 1079.140162] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.150474] env[61867]: DEBUG oslo_vmware.api [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277179, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.153460] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277178, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.331973] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 5ec9eb1f-1d24-41d9-b5c7-a478a427c563] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.424793] env[61867]: DEBUG oslo_concurrency.lockutils [None req-8ba413b3-6ecd-450e-8d9d-0e9bae323aa8 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "6fa6c260-9c20-4808-9dd7-946a27651681" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.208s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.652443] env[61867]: DEBUG oslo_vmware.api [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277179, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186585} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.655440] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1079.655634] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1079.655817] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1079.655994] env[61867]: INFO nova.compute.manager [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1079.656239] env[61867]: DEBUG oslo.service.loopingcall [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1079.656446] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277178, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.656682] env[61867]: DEBUG nova.compute.manager [-] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1079.656778] env[61867]: DEBUG nova.network.neutron [-] [instance: 70170813-af35-4d7f-88e6-2adee247a701] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1079.838580] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 816ca796-d8ed-4843-9b1f-f169f48ff047] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.861068] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e42512-2d16-4b0e-86e3-56404460a454 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.869657] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b8c58e-6137-40d8-b940-4a3b4c2ba14c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.901632] env[61867]: DEBUG nova.network.neutron [-] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.903915] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace557ee-560d-47c0-a7c4-59406af838fa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.907656] env[61867]: DEBUG nova.compute.manager [req-aaa80a59-6d36-49b8-977a-f9618640b8ad req-4f5e3351-b469-4166-8f34-fc70ef8bec4c service nova] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Received event network-vif-deleted-b7b25937-4b43-48af-88a2-628f656bbf4d {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1079.907859] env[61867]: INFO nova.compute.manager [req-aaa80a59-6d36-49b8-977a-f9618640b8ad req-4f5e3351-b469-4166-8f34-fc70ef8bec4c service nova] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Neutron deleted interface b7b25937-4b43-48af-88a2-628f656bbf4d; detaching it from the instance and deleting it from the info cache [ 1079.908309] env[61867]: DEBUG nova.network.neutron [req-aaa80a59-6d36-49b8-977a-f9618640b8ad req-4f5e3351-b469-4166-8f34-fc70ef8bec4c service nova] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.915279] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d8b9b00-baa6-463c-b166-dd3c9f49c120 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.929688] env[61867]: DEBUG nova.compute.provider_tree [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.978321] env[61867]: DEBUG oslo_concurrency.lockutils [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "691b1a16-8cc4-4704-9041-ef3f6f5475e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.978594] env[61867]: DEBUG oslo_concurrency.lockutils [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "691b1a16-8cc4-4704-9041-ef3f6f5475e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.154403] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277178, 'name': ReconfigVM_Task, 'duration_secs': 0.910298} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.154667] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Reconfigured VM instance instance-00000064 to attach disk [datastore1] df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2/df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1080.155523] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90916901-8fd0-4aac-9303-b5d23e53574d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.175150] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594e98cd-9b58-4a94-bb94-c98d50687f51 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.194053] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d8813f-6567-46c3-b10e-bc2078ad3c66 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.214257] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-430d4a01-01ba-4d7c-bef3-66c97742d105 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.220450] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1080.220672] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-efa36aed-5148-4a19-855b-bba4b4f1d1f2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.226313] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1080.226313] env[61867]: value = "task-1277180" [ 1080.226313] env[61867]: _type = "Task" [ 1080.226313] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.234101] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277180, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.341643] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 705a32b6-67f9-42cc-b4d0-f6d1783c68b5] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.410707] env[61867]: INFO nova.compute.manager [-] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Took 0.75 seconds to deallocate network for instance. [ 1080.412841] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5ccc285b-f7c7-4db1-9873-c16874f09bfe {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.424530] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda51dba-426c-4f16-ad97-01178199d216 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.435525] env[61867]: DEBUG nova.scheduler.client.report [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1080.452489] env[61867]: DEBUG nova.compute.manager [req-aaa80a59-6d36-49b8-977a-f9618640b8ad req-4f5e3351-b469-4166-8f34-fc70ef8bec4c service nova] [instance: 70170813-af35-4d7f-88e6-2adee247a701] Detach interface failed, port_id=b7b25937-4b43-48af-88a2-628f656bbf4d, reason: Instance 70170813-af35-4d7f-88e6-2adee247a701 could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1080.480364] env[61867]: DEBUG nova.compute.manager [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1080.736337] env[61867]: DEBUG oslo_vmware.api [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277180, 'name': PowerOnVM_Task, 'duration_secs': 0.435483} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.736563] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1080.844740] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: adc2732d-2a10-40ce-bb90-ed0762a36614] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.920355] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.939802] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.171s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.940367] env[61867]: DEBUG nova.compute.manager [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1080.943460] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.023s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.943709] env[61867]: DEBUG nova.objects.instance [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lazy-loading 'resources' on Instance uuid 70170813-af35-4d7f-88e6-2adee247a701 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1081.002223] env[61867]: DEBUG oslo_concurrency.lockutils [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.348323] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: b8e561df-069e-4873-91c3-1bfe5f27de2d] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.447070] env[61867]: DEBUG nova.compute.utils [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1081.451690] env[61867]: DEBUG nova.compute.manager [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1081.451864] env[61867]: DEBUG nova.network.neutron [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1081.498829] env[61867]: DEBUG nova.policy [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '322fe9fc12a34183b4241e28f7254df4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24180ce8aca142fb897e29d4853c20c3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 1081.532789] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a814e7b3-0208-402f-bcd3-d9ea2a2addcd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.540425] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375d0d21-4bf9-4193-a2a7-5d7b45831e10 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.573273] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e08f76-2dba-4222-b68b-69c3d06e4ba5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.580809] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1c5ccf-205d-4a4a-9f76-09f092e45634 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.594924] env[61867]: DEBUG nova.compute.provider_tree [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.769203] env[61867]: DEBUG nova.network.neutron [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Successfully created port: c7f40157-961b-4250-a4b7-05b534b3d69b {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1081.791417] env[61867]: INFO nova.compute.manager [None req-13df49bd-cc03-403c-a1cb-5b3142a34638 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating instance to original state: 'active' [ 1081.851776] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 7479bf91-5aef-4e75-a127-7e82ae15a003] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.952202] env[61867]: DEBUG nova.compute.manager [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1082.098215] env[61867]: DEBUG nova.scheduler.client.report [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1082.354785] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 07fdd75c-34ab-45e5-a98b-aced2caa6cb5] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.603421] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.660s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.606298] env[61867]: DEBUG oslo_concurrency.lockutils [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.604s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.607538] env[61867]: INFO nova.compute.claims [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1082.624853] env[61867]: INFO nova.scheduler.client.report [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Deleted allocations for instance 70170813-af35-4d7f-88e6-2adee247a701 [ 1082.857730] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 9c23a44c-eb72-4194-a3e5-88a8ef54ed24] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.963683] env[61867]: DEBUG nova.compute.manager [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1082.989407] env[61867]: DEBUG nova.virt.hardware [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1082.989705] env[61867]: DEBUG nova.virt.hardware [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1082.989876] env[61867]: DEBUG nova.virt.hardware [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1082.990102] env[61867]: DEBUG nova.virt.hardware [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1082.990265] env[61867]: DEBUG nova.virt.hardware [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1082.990438] env[61867]: DEBUG nova.virt.hardware [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1082.990661] env[61867]: DEBUG nova.virt.hardware [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1082.990838] env[61867]: DEBUG nova.virt.hardware [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1082.991027] env[61867]: DEBUG nova.virt.hardware [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1082.991212] env[61867]: DEBUG nova.virt.hardware [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1082.991442] env[61867]: DEBUG nova.virt.hardware [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1082.992362] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc864511-f85b-40b7-9060-ee65ab58f1dc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.000558] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b880d4-413a-4e0a-85f5-b26c8bafc3a3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.131725] env[61867]: DEBUG oslo_concurrency.lockutils [None req-f1abd1e6-e1f8-4bd9-ab4d-2bbad042be8b tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "70170813-af35-4d7f-88e6-2adee247a701" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.638s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.195521] env[61867]: DEBUG nova.compute.manager [req-7d97e1d0-ff5a-493b-8654-d599a3d182a9 req-d8450138-5b65-47dc-9769-563a1085c7a8 service nova] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Received event network-vif-plugged-c7f40157-961b-4250-a4b7-05b534b3d69b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1083.195762] env[61867]: DEBUG oslo_concurrency.lockutils [req-7d97e1d0-ff5a-493b-8654-d599a3d182a9 req-d8450138-5b65-47dc-9769-563a1085c7a8 service nova] Acquiring lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.195977] env[61867]: DEBUG oslo_concurrency.lockutils [req-7d97e1d0-ff5a-493b-8654-d599a3d182a9 req-d8450138-5b65-47dc-9769-563a1085c7a8 service nova] Lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.196166] env[61867]: DEBUG oslo_concurrency.lockutils [req-7d97e1d0-ff5a-493b-8654-d599a3d182a9 req-d8450138-5b65-47dc-9769-563a1085c7a8 service nova] Lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.196653] env[61867]: DEBUG nova.compute.manager [req-7d97e1d0-ff5a-493b-8654-d599a3d182a9 req-d8450138-5b65-47dc-9769-563a1085c7a8 service nova] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] No waiting events found dispatching network-vif-plugged-c7f40157-961b-4250-a4b7-05b534b3d69b {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1083.196843] env[61867]: WARNING nova.compute.manager [req-7d97e1d0-ff5a-493b-8654-d599a3d182a9 req-d8450138-5b65-47dc-9769-563a1085c7a8 service nova] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Received unexpected event network-vif-plugged-c7f40157-961b-4250-a4b7-05b534b3d69b for instance with vm_state building and task_state spawning. [ 1083.360969] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 2aa08603-d87f-4734-bdfe-fdd610d54e1f] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.556863] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.557186] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.557423] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.557650] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.557830] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.559691] env[61867]: INFO nova.compute.manager [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Terminating instance [ 1083.561544] env[61867]: DEBUG nova.compute.manager [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1083.561764] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1083.562006] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db6787bf-aa05-4f0f-9601-b6abe846fc01 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.569567] env[61867]: DEBUG oslo_vmware.api [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1083.569567] env[61867]: value = "task-1277181" [ 1083.569567] env[61867]: _type = "Task" [ 1083.569567] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.583314] env[61867]: DEBUG oslo_vmware.api [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277181, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.694961] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f280d3c0-7336-486c-b3c3-ed65902d76c3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.703039] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c871be3a-fe8d-40a9-88bf-358d31da15bd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.732628] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab3c6baa-b1d6-46a1-b052-50da5a617a0b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.736934] env[61867]: DEBUG nova.network.neutron [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Successfully updated port: c7f40157-961b-4250-a4b7-05b534b3d69b {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1083.741202] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faeb4e36-f638-479c-bcdc-cbd75cdc7121 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.760241] env[61867]: DEBUG nova.compute.provider_tree [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.772257] env[61867]: DEBUG nova.compute.manager [req-c48a4301-422d-43da-b63e-230abb17a283 req-96962034-15cd-48a5-b353-9b26d6ca459e service nova] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Received event network-changed-c7f40157-961b-4250-a4b7-05b534b3d69b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1083.772481] env[61867]: DEBUG nova.compute.manager [req-c48a4301-422d-43da-b63e-230abb17a283 req-96962034-15cd-48a5-b353-9b26d6ca459e service nova] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Refreshing instance network info cache due to event network-changed-c7f40157-961b-4250-a4b7-05b534b3d69b. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1083.772687] env[61867]: DEBUG oslo_concurrency.lockutils [req-c48a4301-422d-43da-b63e-230abb17a283 req-96962034-15cd-48a5-b353-9b26d6ca459e service nova] Acquiring lock "refresh_cache-0b54931d-d5cd-459f-a6a2-e941c4f360d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.772817] env[61867]: DEBUG oslo_concurrency.lockutils [req-c48a4301-422d-43da-b63e-230abb17a283 req-96962034-15cd-48a5-b353-9b26d6ca459e service nova] Acquired lock "refresh_cache-0b54931d-d5cd-459f-a6a2-e941c4f360d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.772978] env[61867]: DEBUG nova.network.neutron [req-c48a4301-422d-43da-b63e-230abb17a283 req-96962034-15cd-48a5-b353-9b26d6ca459e service nova] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Refreshing network info cache for port c7f40157-961b-4250-a4b7-05b534b3d69b {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1083.864394] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 6d2dab88-4165-4952-8019-2eaf3b863115] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.038944] env[61867]: DEBUG oslo_concurrency.lockutils [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "a9724422-abd1-4771-acb2-2c0ac45f5c16" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.039335] env[61867]: DEBUG oslo_concurrency.lockutils [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "a9724422-abd1-4771-acb2-2c0ac45f5c16" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.039659] env[61867]: DEBUG oslo_concurrency.lockutils [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "a9724422-abd1-4771-acb2-2c0ac45f5c16-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.039955] env[61867]: DEBUG oslo_concurrency.lockutils [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "a9724422-abd1-4771-acb2-2c0ac45f5c16-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.040237] env[61867]: DEBUG oslo_concurrency.lockutils [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "a9724422-abd1-4771-acb2-2c0ac45f5c16-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.043602] env[61867]: INFO nova.compute.manager [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Terminating instance [ 1084.046224] env[61867]: DEBUG nova.compute.manager [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1084.046436] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1084.047571] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b9dcc37-f0be-4547-bcaa-145ba7988a42 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.054912] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1084.055170] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-258ac3fc-45b0-4ef7-be4a-8c69fb472905 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.061577] env[61867]: DEBUG oslo_vmware.api [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1084.061577] env[61867]: value = "task-1277182" [ 1084.061577] env[61867]: _type = "Task" [ 1084.061577] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.072824] env[61867]: DEBUG oslo_vmware.api [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277182, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.081609] env[61867]: DEBUG oslo_vmware.api [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277181, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.247274] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "refresh_cache-0b54931d-d5cd-459f-a6a2-e941c4f360d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.264049] env[61867]: DEBUG nova.scheduler.client.report [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1084.304690] env[61867]: DEBUG nova.network.neutron [req-c48a4301-422d-43da-b63e-230abb17a283 req-96962034-15cd-48a5-b353-9b26d6ca459e service nova] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1084.368661] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 71cd9036-0f99-4e30-aad6-ceb4f15d4ffe] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.375159] env[61867]: DEBUG nova.network.neutron [req-c48a4301-422d-43da-b63e-230abb17a283 req-96962034-15cd-48a5-b353-9b26d6ca459e service nova] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.571500] env[61867]: DEBUG oslo_vmware.api [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277182, 'name': PowerOffVM_Task, 'duration_secs': 0.155724} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.571717] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1084.571879] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1084.572137] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b3fce25-f0f3-47f1-8580-dc52895618d2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.581199] env[61867]: DEBUG oslo_vmware.api [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277181, 'name': PowerOffVM_Task, 'duration_secs': 0.529448} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.581432] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1084.581630] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Volume detach. Driver type: vmdk {{(pid=61867) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1084.581819] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274421', 'volume_id': 'e8d5686e-270a-4ffc-9ba7-aa3f11f81510', 'name': 'volume-e8d5686e-270a-4ffc-9ba7-aa3f11f81510', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2', 'attached_at': '2024-10-12T12:53:15.000000', 'detached_at': '', 'volume_id': 'e8d5686e-270a-4ffc-9ba7-aa3f11f81510', 'serial': 'e8d5686e-270a-4ffc-9ba7-aa3f11f81510'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1084.582521] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba1f7de-64c3-4db4-97a7-cae7dcf7679c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.603734] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e0e525-c925-43ef-a489-d69eae2cee5d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.610482] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c9c9253-4a42-4e17-a349-eec5c6815367 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.630327] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8dac4c3-2160-4530-8082-a893c9da498a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.644703] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] The volume has not been displaced from its original location: [datastore2] volume-e8d5686e-270a-4ffc-9ba7-aa3f11f81510/volume-e8d5686e-270a-4ffc-9ba7-aa3f11f81510.vmdk. No consolidation needed. {{(pid=61867) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1084.650039] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Reconfiguring VM instance instance-00000064 to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1084.651242] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-782d9095-5724-4324-94d7-f3fac2dbfe97 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.663939] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1084.664158] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1084.664341] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Deleting the datastore file [datastore2] a9724422-abd1-4771-acb2-2c0ac45f5c16 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1084.664590] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6cd551a7-121e-4381-9c0a-5ddaacb62152 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.672260] env[61867]: DEBUG oslo_vmware.api [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1084.672260] env[61867]: value = "task-1277185" [ 1084.672260] env[61867]: _type = "Task" [ 1084.672260] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.673357] env[61867]: DEBUG oslo_vmware.api [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for the task: (returnval){ [ 1084.673357] env[61867]: value = "task-1277184" [ 1084.673357] env[61867]: _type = "Task" [ 1084.673357] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.683533] env[61867]: DEBUG oslo_vmware.api [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277184, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.686391] env[61867]: DEBUG oslo_vmware.api [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277185, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.768477] env[61867]: DEBUG oslo_concurrency.lockutils [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.162s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.769046] env[61867]: DEBUG nova.compute.manager [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1084.871749] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: e689c3c3-cdb8-40fd-9b65-c9cc6ed67d2b] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.877884] env[61867]: DEBUG oslo_concurrency.lockutils [req-c48a4301-422d-43da-b63e-230abb17a283 req-96962034-15cd-48a5-b353-9b26d6ca459e service nova] Releasing lock "refresh_cache-0b54931d-d5cd-459f-a6a2-e941c4f360d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.879259] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquired lock "refresh_cache-0b54931d-d5cd-459f-a6a2-e941c4f360d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.879259] env[61867]: DEBUG nova.network.neutron [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1085.185674] env[61867]: DEBUG oslo_vmware.api [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277185, 'name': ReconfigVM_Task, 'duration_secs': 0.193532} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.188540] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Reconfigured VM instance instance-00000064 to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1085.193066] env[61867]: DEBUG oslo_vmware.api [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Task: {'id': task-1277184, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178515} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.193299] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c7c17ba-9700-4cd0-855d-003db93d5f59 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.202962] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1085.203207] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1085.203356] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1085.203572] env[61867]: INFO nova.compute.manager [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1085.203815] env[61867]: DEBUG oslo.service.loopingcall [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1085.204033] env[61867]: DEBUG nova.compute.manager [-] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1085.204137] env[61867]: DEBUG nova.network.neutron [-] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1085.210309] env[61867]: DEBUG oslo_vmware.api [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1085.210309] env[61867]: value = "task-1277186" [ 1085.210309] env[61867]: _type = "Task" [ 1085.210309] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.217993] env[61867]: DEBUG oslo_vmware.api [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277186, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.274206] env[61867]: DEBUG nova.compute.utils [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1085.275666] env[61867]: DEBUG nova.compute.manager [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1085.275855] env[61867]: DEBUG nova.network.neutron [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1085.339449] env[61867]: DEBUG nova.policy [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bcdb4a387cdb45899eded851c1bc842a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f245b07249fa4e37ac97cf66dbf6678f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 1085.375443] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 410a2819-ea27-4613-9f2c-279f9e82a4b3] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.440953] env[61867]: DEBUG nova.network.neutron [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1085.721347] env[61867]: DEBUG oslo_vmware.api [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277186, 'name': ReconfigVM_Task, 'duration_secs': 0.154418} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.721713] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274421', 'volume_id': 'e8d5686e-270a-4ffc-9ba7-aa3f11f81510', 'name': 'volume-e8d5686e-270a-4ffc-9ba7-aa3f11f81510', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2', 'attached_at': '2024-10-12T12:53:15.000000', 'detached_at': '', 'volume_id': 'e8d5686e-270a-4ffc-9ba7-aa3f11f81510', 'serial': 'e8d5686e-270a-4ffc-9ba7-aa3f11f81510'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1085.722065] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1085.722883] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3208357-0ffe-460c-aafe-dc3cc7bf6d63 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.726781] env[61867]: DEBUG nova.network.neutron [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Updating instance_info_cache with network_info: [{"id": "c7f40157-961b-4250-a4b7-05b534b3d69b", "address": "fa:16:3e:51:60:b1", "network": {"id": "fc403caf-421f-4783-9807-5e61c4379a6d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-846804042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24180ce8aca142fb897e29d4853c20c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7f40157-96", "ovs_interfaceid": "c7f40157-961b-4250-a4b7-05b534b3d69b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.729714] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1085.730183] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-19abc762-daff-4d8b-883f-efb9ac611f67 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.778896] env[61867]: DEBUG nova.compute.manager [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1085.789034] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1085.789140] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1085.789302] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Deleting the datastore file [datastore1] df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1085.789571] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b9f3a345-d251-4ac4-af3b-83cf641ba00c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.796130] env[61867]: DEBUG oslo_vmware.api [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1085.796130] env[61867]: value = "task-1277188" [ 1085.796130] env[61867]: _type = "Task" [ 1085.796130] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.801470] env[61867]: DEBUG nova.compute.manager [req-17c2e5e3-6960-409d-a499-4663d6b2ec47 req-1eccc69f-28b1-4e74-8777-607d1ca1c086 service nova] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Received event network-vif-deleted-ddcbc24f-ddfe-46d9-8ad4-214cd357022d {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1085.802228] env[61867]: INFO nova.compute.manager [req-17c2e5e3-6960-409d-a499-4663d6b2ec47 req-1eccc69f-28b1-4e74-8777-607d1ca1c086 service nova] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Neutron deleted interface ddcbc24f-ddfe-46d9-8ad4-214cd357022d; detaching it from the instance and deleting it from the info cache [ 1085.802228] env[61867]: DEBUG nova.network.neutron [req-17c2e5e3-6960-409d-a499-4663d6b2ec47 req-1eccc69f-28b1-4e74-8777-607d1ca1c086 service nova] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.808441] env[61867]: DEBUG oslo_vmware.api [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277188, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.812077] env[61867]: DEBUG nova.network.neutron [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Successfully created port: 218c1590-49c5-460e-b6bf-03377d5a3c75 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1085.879765] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: ca2f6d18-f773-4875-83de-2f2be912c2f8] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.135118] env[61867]: DEBUG nova.network.neutron [-] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.231587] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Releasing lock "refresh_cache-0b54931d-d5cd-459f-a6a2-e941c4f360d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.233659] env[61867]: DEBUG nova.compute.manager [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Instance network_info: |[{"id": "c7f40157-961b-4250-a4b7-05b534b3d69b", "address": "fa:16:3e:51:60:b1", "network": {"id": "fc403caf-421f-4783-9807-5e61c4379a6d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-846804042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24180ce8aca142fb897e29d4853c20c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7f40157-96", "ovs_interfaceid": "c7f40157-961b-4250-a4b7-05b534b3d69b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1086.233659] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:51:60:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '47ca1ce6-8148-48d5-bcfe-89e39b73914e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c7f40157-961b-4250-a4b7-05b534b3d69b', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1086.240370] env[61867]: DEBUG oslo.service.loopingcall [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1086.241128] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1086.241128] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c4dfa654-646c-417a-a451-1041be59daa4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.260776] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1086.260776] env[61867]: value = "task-1277189" [ 1086.260776] env[61867]: _type = "Task" [ 1086.260776] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.269068] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277189, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.306757] env[61867]: DEBUG oslo_vmware.api [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277188, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187294} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.307453] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d289102c-4edf-465a-a9e6-25c7264e8145 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.309082] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1086.309277] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1086.309485] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1086.309705] env[61867]: INFO nova.compute.manager [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Took 2.75 seconds to destroy the instance on the hypervisor. [ 1086.310662] env[61867]: DEBUG oslo.service.loopingcall [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1086.310662] env[61867]: DEBUG nova.compute.manager [-] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1086.310662] env[61867]: DEBUG nova.network.neutron [-] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1086.318838] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a32c6be-d0f4-452f-ba47-30850d5b6f13 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.346412] env[61867]: DEBUG nova.compute.manager [req-17c2e5e3-6960-409d-a499-4663d6b2ec47 req-1eccc69f-28b1-4e74-8777-607d1ca1c086 service nova] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Detach interface failed, port_id=ddcbc24f-ddfe-46d9-8ad4-214cd357022d, reason: Instance a9724422-abd1-4771-acb2-2c0ac45f5c16 could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1086.382746] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: ad699b24-d01d-4d7a-815f-c6b10286012d] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.637130] env[61867]: INFO nova.compute.manager [-] [instance: a9724422-abd1-4771-acb2-2c0ac45f5c16] Took 1.43 seconds to deallocate network for instance. [ 1086.771377] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277189, 'name': CreateVM_Task, 'duration_secs': 0.412121} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.771555] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1086.772272] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.772443] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.772779] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1086.773052] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a439243b-7957-48cf-80bc-f76e31c49a7e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.778082] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1086.778082] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f8e324-5700-ef86-558e-c06b5a8e794e" [ 1086.778082] env[61867]: _type = "Task" [ 1086.778082] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.785800] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f8e324-5700-ef86-558e-c06b5a8e794e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.786833] env[61867]: DEBUG nova.compute.manager [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1086.818546] env[61867]: DEBUG nova.virt.hardware [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1086.818796] env[61867]: DEBUG nova.virt.hardware [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1086.818955] env[61867]: DEBUG nova.virt.hardware [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1086.819159] env[61867]: DEBUG nova.virt.hardware [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1086.819311] env[61867]: DEBUG nova.virt.hardware [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1086.819464] env[61867]: DEBUG nova.virt.hardware [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1086.819676] env[61867]: DEBUG nova.virt.hardware [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1086.819836] env[61867]: DEBUG nova.virt.hardware [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1086.820010] env[61867]: DEBUG nova.virt.hardware [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1086.820186] env[61867]: DEBUG nova.virt.hardware [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1086.820362] env[61867]: DEBUG nova.virt.hardware [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1086.821299] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b522460-d448-4b77-988d-0eed75a4a993 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.831958] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f419ac65-5a21-4fe7-88db-5c91121525e2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.886038] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 9aab8852-addb-49e6-a59b-fa9bffc7733b] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.144649] env[61867]: DEBUG oslo_concurrency.lockutils [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.145064] env[61867]: DEBUG oslo_concurrency.lockutils [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.145360] env[61867]: DEBUG nova.objects.instance [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lazy-loading 'resources' on Instance uuid a9724422-abd1-4771-acb2-2c0ac45f5c16 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.289599] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f8e324-5700-ef86-558e-c06b5a8e794e, 'name': SearchDatastore_Task, 'duration_secs': 0.025873} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.289948] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.290285] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1087.290558] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.290710] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.290892] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1087.291186] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-505fd0a4-dba1-440e-8e41-f4b1f978f33c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.301452] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1087.301688] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1087.302566] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c667d32e-df99-4763-b9bc-1af7a32fefb7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.308506] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1087.308506] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5243e62a-6bd5-3ce3-a773-932b265df482" [ 1087.308506] env[61867]: _type = "Task" [ 1087.308506] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.316226] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5243e62a-6bd5-3ce3-a773-932b265df482, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.326036] env[61867]: DEBUG nova.network.neutron [-] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.342360] env[61867]: DEBUG nova.network.neutron [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Successfully updated port: 218c1590-49c5-460e-b6bf-03377d5a3c75 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1087.389365] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 6dda389d-a8c5-4e0e-87a5-4065e24c034e] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.716308] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb1c750c-9548-4eed-8392-2a82d2e1e91d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.723905] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e8bc48-ab7d-4301-bf61-142ab1b316a9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.752595] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8974a95-aa4c-4b68-bfde-f6fc55309cb2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.759548] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3bf0678-5bb6-4678-9fd0-8e5fb3bb757a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.773658] env[61867]: DEBUG nova.compute.provider_tree [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1087.817568] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5243e62a-6bd5-3ce3-a773-932b265df482, 'name': SearchDatastore_Task, 'duration_secs': 0.022819} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.818459] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a2b49ef-f3f1-4d0b-8ac0-f3ce870d851d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.823168] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1087.823168] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e4a41f-26f4-db06-2cc5-7e0784fcb67c" [ 1087.823168] env[61867]: _type = "Task" [ 1087.823168] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.830550] env[61867]: INFO nova.compute.manager [-] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Took 1.52 seconds to deallocate network for instance. [ 1087.830826] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e4a41f-26f4-db06-2cc5-7e0784fcb67c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.834779] env[61867]: DEBUG nova.compute.manager [req-abf755ac-ff56-4fe6-a9af-5e4fd2860933 req-2de936ad-0507-4216-9b99-b9bf31c57ffd service nova] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Received event network-vif-deleted-f1e84fbf-9fe3-4e2b-a89b-e5c05b148fc2 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1087.834979] env[61867]: DEBUG nova.compute.manager [req-abf755ac-ff56-4fe6-a9af-5e4fd2860933 req-2de936ad-0507-4216-9b99-b9bf31c57ffd service nova] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Received event network-vif-plugged-218c1590-49c5-460e-b6bf-03377d5a3c75 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1087.835210] env[61867]: DEBUG oslo_concurrency.lockutils [req-abf755ac-ff56-4fe6-a9af-5e4fd2860933 req-2de936ad-0507-4216-9b99-b9bf31c57ffd service nova] Acquiring lock "691b1a16-8cc4-4704-9041-ef3f6f5475e2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.835376] env[61867]: DEBUG oslo_concurrency.lockutils [req-abf755ac-ff56-4fe6-a9af-5e4fd2860933 req-2de936ad-0507-4216-9b99-b9bf31c57ffd service nova] Lock "691b1a16-8cc4-4704-9041-ef3f6f5475e2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.835544] env[61867]: DEBUG oslo_concurrency.lockutils [req-abf755ac-ff56-4fe6-a9af-5e4fd2860933 req-2de936ad-0507-4216-9b99-b9bf31c57ffd service nova] Lock "691b1a16-8cc4-4704-9041-ef3f6f5475e2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.835710] env[61867]: DEBUG nova.compute.manager [req-abf755ac-ff56-4fe6-a9af-5e4fd2860933 req-2de936ad-0507-4216-9b99-b9bf31c57ffd service nova] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] No waiting events found dispatching network-vif-plugged-218c1590-49c5-460e-b6bf-03377d5a3c75 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1087.835882] env[61867]: WARNING nova.compute.manager [req-abf755ac-ff56-4fe6-a9af-5e4fd2860933 req-2de936ad-0507-4216-9b99-b9bf31c57ffd service nova] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Received unexpected event network-vif-plugged-218c1590-49c5-460e-b6bf-03377d5a3c75 for instance with vm_state building and task_state spawning. [ 1087.836048] env[61867]: DEBUG nova.compute.manager [req-abf755ac-ff56-4fe6-a9af-5e4fd2860933 req-2de936ad-0507-4216-9b99-b9bf31c57ffd service nova] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Received event network-changed-218c1590-49c5-460e-b6bf-03377d5a3c75 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1087.836205] env[61867]: DEBUG nova.compute.manager [req-abf755ac-ff56-4fe6-a9af-5e4fd2860933 req-2de936ad-0507-4216-9b99-b9bf31c57ffd service nova] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Refreshing instance network info cache due to event network-changed-218c1590-49c5-460e-b6bf-03377d5a3c75. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1087.836386] env[61867]: DEBUG oslo_concurrency.lockutils [req-abf755ac-ff56-4fe6-a9af-5e4fd2860933 req-2de936ad-0507-4216-9b99-b9bf31c57ffd service nova] Acquiring lock "refresh_cache-691b1a16-8cc4-4704-9041-ef3f6f5475e2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.836522] env[61867]: DEBUG oslo_concurrency.lockutils [req-abf755ac-ff56-4fe6-a9af-5e4fd2860933 req-2de936ad-0507-4216-9b99-b9bf31c57ffd service nova] Acquired lock "refresh_cache-691b1a16-8cc4-4704-9041-ef3f6f5475e2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.836712] env[61867]: DEBUG nova.network.neutron [req-abf755ac-ff56-4fe6-a9af-5e4fd2860933 req-2de936ad-0507-4216-9b99-b9bf31c57ffd service nova] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Refreshing network info cache for port 218c1590-49c5-460e-b6bf-03377d5a3c75 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1087.847214] env[61867]: DEBUG oslo_concurrency.lockutils [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "refresh_cache-691b1a16-8cc4-4704-9041-ef3f6f5475e2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.892602] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 8a83f4f2-58eb-473e-9b1e-32ce633554f9] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.276811] env[61867]: DEBUG nova.scheduler.client.report [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1088.333264] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e4a41f-26f4-db06-2cc5-7e0784fcb67c, 'name': SearchDatastore_Task, 'duration_secs': 0.009704} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.333570] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.333837] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 0b54931d-d5cd-459f-a6a2-e941c4f360d6/0b54931d-d5cd-459f-a6a2-e941c4f360d6.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1088.334126] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-376e98b9-e584-483e-939e-4d488c636be9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.341213] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1088.341213] env[61867]: value = "task-1277190" [ 1088.341213] env[61867]: _type = "Task" [ 1088.341213] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.349390] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277190, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.370513] env[61867]: DEBUG nova.network.neutron [req-abf755ac-ff56-4fe6-a9af-5e4fd2860933 req-2de936ad-0507-4216-9b99-b9bf31c57ffd service nova] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1088.375060] env[61867]: INFO nova.compute.manager [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2] Took 0.54 seconds to detach 1 volumes for instance. [ 1088.396014] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 6e41989e-b8fa-4009-af1e-1ce859b329a1] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.445318] env[61867]: DEBUG nova.network.neutron [req-abf755ac-ff56-4fe6-a9af-5e4fd2860933 req-2de936ad-0507-4216-9b99-b9bf31c57ffd service nova] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.781641] env[61867]: DEBUG oslo_concurrency.lockutils [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.636s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.800911] env[61867]: INFO nova.scheduler.client.report [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Deleted allocations for instance a9724422-abd1-4771-acb2-2c0ac45f5c16 [ 1088.851565] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277190, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.427589} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.851840] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 0b54931d-d5cd-459f-a6a2-e941c4f360d6/0b54931d-d5cd-459f-a6a2-e941c4f360d6.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1088.852067] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1088.852313] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5ad4fb81-bc0e-406a-b212-c4e24b82a8fd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.858317] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1088.858317] env[61867]: value = "task-1277191" [ 1088.858317] env[61867]: _type = "Task" [ 1088.858317] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.865519] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277191, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.880880] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.881105] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.881301] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.897726] env[61867]: INFO nova.scheduler.client.report [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Deleted allocations for instance df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2 [ 1088.898854] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: abb41c0c-6d0d-4147-a4af-554ab7d9e921] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.947416] env[61867]: DEBUG oslo_concurrency.lockutils [req-abf755ac-ff56-4fe6-a9af-5e4fd2860933 req-2de936ad-0507-4216-9b99-b9bf31c57ffd service nova] Releasing lock "refresh_cache-691b1a16-8cc4-4704-9041-ef3f6f5475e2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.947790] env[61867]: DEBUG oslo_concurrency.lockutils [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquired lock "refresh_cache-691b1a16-8cc4-4704-9041-ef3f6f5475e2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.947944] env[61867]: DEBUG nova.network.neutron [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1089.309841] env[61867]: DEBUG oslo_concurrency.lockutils [None req-52955c2f-e00e-48fb-a0d7-608c7cb470e8 tempest-ServerRescueTestJSON-209360935 tempest-ServerRescueTestJSON-209360935-project-member] Lock "a9724422-abd1-4771-acb2-2c0ac45f5c16" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.270s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.369093] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277191, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05764} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.369093] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1089.369465] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7192f93d-39f8-49bb-8834-c85ec685d711 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.392628] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 0b54931d-d5cd-459f-a6a2-e941c4f360d6/0b54931d-d5cd-459f-a6a2-e941c4f360d6.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1089.393568] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e45427c7-7350-450a-bc43-1b9628f8692c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.409148] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 66518b2a-0242-438b-ba9f-d57c07a1165c] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.414953] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9dd8d82d-2479-4026-af20-96ec3a7bded2 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "df94dc6d-a72a-4c6d-b3aa-0cd7a1be67c2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.858s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.420017] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1089.420017] env[61867]: value = "task-1277192" [ 1089.420017] env[61867]: _type = "Task" [ 1089.420017] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.429409] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277192, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.477233] env[61867]: DEBUG nova.network.neutron [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1089.603340] env[61867]: DEBUG nova.network.neutron [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Updating instance_info_cache with network_info: [{"id": "218c1590-49c5-460e-b6bf-03377d5a3c75", "address": "fa:16:3e:5f:ef:82", "network": {"id": "8db0c704-eb45-4fca-8624-d21877ab5ab3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-74001547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f245b07249fa4e37ac97cf66dbf6678f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap218c1590-49", "ovs_interfaceid": "218c1590-49c5-460e-b6bf-03377d5a3c75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.913538] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 4653ce96-257e-46d8-8c3d-85c03380213c] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.929972] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277192, 'name': ReconfigVM_Task, 'duration_secs': 0.259133} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.930482] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 0b54931d-d5cd-459f-a6a2-e941c4f360d6/0b54931d-d5cd-459f-a6a2-e941c4f360d6.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1089.931147] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d7924396-592b-4458-bb85-599d51c57027 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.938956] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1089.938956] env[61867]: value = "task-1277193" [ 1089.938956] env[61867]: _type = "Task" [ 1089.938956] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.947920] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277193, 'name': Rename_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.106528] env[61867]: DEBUG oslo_concurrency.lockutils [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Releasing lock "refresh_cache-691b1a16-8cc4-4704-9041-ef3f6f5475e2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.107108] env[61867]: DEBUG nova.compute.manager [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Instance network_info: |[{"id": "218c1590-49c5-460e-b6bf-03377d5a3c75", "address": "fa:16:3e:5f:ef:82", "network": {"id": "8db0c704-eb45-4fca-8624-d21877ab5ab3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-74001547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f245b07249fa4e37ac97cf66dbf6678f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap218c1590-49", "ovs_interfaceid": "218c1590-49c5-460e-b6bf-03377d5a3c75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1090.107502] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5f:ef:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91c1da19-ab68-4127-bacd-accbaff19651', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '218c1590-49c5-460e-b6bf-03377d5a3c75', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1090.116786] env[61867]: DEBUG oslo.service.loopingcall [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1090.117114] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1090.117551] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-698ea4a5-a465-43ba-9619-9b03125ceb35 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.141609] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1090.141609] env[61867]: value = "task-1277194" [ 1090.141609] env[61867]: _type = "Task" [ 1090.141609] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.148742] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277194, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.417215] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 81e70c36-04ea-450c-9383-53ef069d1c46] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.451341] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277193, 'name': Rename_Task, 'duration_secs': 0.133605} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.451935] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1090.452722] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-252a6fe7-2ad8-4470-a046-af55ee132ea1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.460570] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1090.460570] env[61867]: value = "task-1277195" [ 1090.460570] env[61867]: _type = "Task" [ 1090.460570] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.469599] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277195, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.652591] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277194, 'name': CreateVM_Task, 'duration_secs': 0.349394} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.652858] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1090.653483] env[61867]: DEBUG oslo_concurrency.lockutils [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.653705] env[61867]: DEBUG oslo_concurrency.lockutils [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.654051] env[61867]: DEBUG oslo_concurrency.lockutils [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1090.654319] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41d8a643-f767-4d3a-b125-51ce13425a13 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.658806] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1090.658806] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527a362a-8f94-6aa2-8e8f-97a98c606dbe" [ 1090.658806] env[61867]: _type = "Task" [ 1090.658806] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.667811] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527a362a-8f94-6aa2-8e8f-97a98c606dbe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.920924] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 1df8427c-e75d-4b60-a92a-b5ba76b67081] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.959358] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.959645] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.971174] env[61867]: DEBUG oslo_vmware.api [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277195, 'name': PowerOnVM_Task, 'duration_secs': 0.466752} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.971430] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1090.971627] env[61867]: INFO nova.compute.manager [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Took 8.01 seconds to spawn the instance on the hypervisor. [ 1090.971799] env[61867]: DEBUG nova.compute.manager [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1090.972937] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac2cf52-a5ff-4163-8ce8-c8d8b9df6c1d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.169357] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]527a362a-8f94-6aa2-8e8f-97a98c606dbe, 'name': SearchDatastore_Task, 'duration_secs': 0.011558} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.169689] env[61867]: DEBUG oslo_concurrency.lockutils [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.169926] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1091.170177] env[61867]: DEBUG oslo_concurrency.lockutils [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.170330] env[61867]: DEBUG oslo_concurrency.lockutils [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.170517] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1091.170797] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-85bd7d87-d673-48d8-bfa5-6dff348de0c6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.179185] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1091.179185] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1091.179919] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1512f964-1b42-45f9-9599-f376e304eb1b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.185047] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1091.185047] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5261e8a8-9c4a-db70-7534-128f95eec815" [ 1091.185047] env[61867]: _type = "Task" [ 1091.185047] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.192261] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5261e8a8-9c4a-db70-7534-128f95eec815, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.426852] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 80a703b3-d692-4023-a73b-dba980a94dff] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.464871] env[61867]: DEBUG nova.compute.manager [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1091.497080] env[61867]: INFO nova.compute.manager [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Took 12.74 seconds to build instance. [ 1091.695332] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5261e8a8-9c4a-db70-7534-128f95eec815, 'name': SearchDatastore_Task, 'duration_secs': 0.018488} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.696131] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4bfb6ff-1b71-46b5-85d0-c25359d14e45 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.701575] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1091.701575] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523416b3-2b95-cc9f-8474-dd9f6a65c568" [ 1091.701575] env[61867]: _type = "Task" [ 1091.701575] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.708887] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523416b3-2b95-cc9f-8474-dd9f6a65c568, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.929610] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: be825f62-cb56-4647-be1c-1605536fdc89] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.986892] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.987179] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.988775] env[61867]: INFO nova.compute.claims [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1092.000024] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9fe1830d-6140-4b33-a8e0-026db65804cc tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.257s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.212086] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]523416b3-2b95-cc9f-8474-dd9f6a65c568, 'name': SearchDatastore_Task, 'duration_secs': 0.036892} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.212461] env[61867]: DEBUG oslo_concurrency.lockutils [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.212661] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 691b1a16-8cc4-4704-9041-ef3f6f5475e2/691b1a16-8cc4-4704-9041-ef3f6f5475e2.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1092.212925] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7eaf14a8-5163-48e7-8818-6bf5889fde3a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.220301] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1092.220301] env[61867]: value = "task-1277196" [ 1092.220301] env[61867]: _type = "Task" [ 1092.220301] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.229208] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277196, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.433239] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 3022ea79-7c27-40a4-80d6-1a86b354917a] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.560955] env[61867]: DEBUG nova.compute.manager [req-9dc5cdc8-1fcf-44cb-af23-b451952aa0c4 req-3d89d486-0966-493e-b4f1-0772ad213a9f service nova] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Received event network-changed-c7f40157-961b-4250-a4b7-05b534b3d69b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1092.561147] env[61867]: DEBUG nova.compute.manager [req-9dc5cdc8-1fcf-44cb-af23-b451952aa0c4 req-3d89d486-0966-493e-b4f1-0772ad213a9f service nova] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Refreshing instance network info cache due to event network-changed-c7f40157-961b-4250-a4b7-05b534b3d69b. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1092.561516] env[61867]: DEBUG oslo_concurrency.lockutils [req-9dc5cdc8-1fcf-44cb-af23-b451952aa0c4 req-3d89d486-0966-493e-b4f1-0772ad213a9f service nova] Acquiring lock "refresh_cache-0b54931d-d5cd-459f-a6a2-e941c4f360d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.561699] env[61867]: DEBUG oslo_concurrency.lockutils [req-9dc5cdc8-1fcf-44cb-af23-b451952aa0c4 req-3d89d486-0966-493e-b4f1-0772ad213a9f service nova] Acquired lock "refresh_cache-0b54931d-d5cd-459f-a6a2-e941c4f360d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.561870] env[61867]: DEBUG nova.network.neutron [req-9dc5cdc8-1fcf-44cb-af23-b451952aa0c4 req-3d89d486-0966-493e-b4f1-0772ad213a9f service nova] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Refreshing network info cache for port c7f40157-961b-4250-a4b7-05b534b3d69b {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1092.732979] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277196, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.938018] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 905ee661-d5bd-4121-9bd0-fe68bdabd2fd] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.086930] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92ee2e0b-ae9e-4eca-8b97-2c7cff86728f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.097150] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-433a0bfd-fc1e-4f04-9bc5-b07cad741481 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.143323] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54d1561-d113-480f-b63c-c8ddacdced1c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.153478] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-805982fe-a347-4d1c-bdc8-b43c85e94553 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.167462] env[61867]: DEBUG nova.compute.provider_tree [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.230445] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277196, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.654669} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.230749] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore2] 691b1a16-8cc4-4704-9041-ef3f6f5475e2/691b1a16-8cc4-4704-9041-ef3f6f5475e2.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1093.231318] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1093.231318] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7ead8006-645c-4818-a33f-13bad914ce8e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.239252] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1093.239252] env[61867]: value = "task-1277197" [ 1093.239252] env[61867]: _type = "Task" [ 1093.239252] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.247456] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277197, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.394484] env[61867]: DEBUG nova.network.neutron [req-9dc5cdc8-1fcf-44cb-af23-b451952aa0c4 req-3d89d486-0966-493e-b4f1-0772ad213a9f service nova] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Updated VIF entry in instance network info cache for port c7f40157-961b-4250-a4b7-05b534b3d69b. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1093.395119] env[61867]: DEBUG nova.network.neutron [req-9dc5cdc8-1fcf-44cb-af23-b451952aa0c4 req-3d89d486-0966-493e-b4f1-0772ad213a9f service nova] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Updating instance_info_cache with network_info: [{"id": "c7f40157-961b-4250-a4b7-05b534b3d69b", "address": "fa:16:3e:51:60:b1", "network": {"id": "fc403caf-421f-4783-9807-5e61c4379a6d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-846804042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.134", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24180ce8aca142fb897e29d4853c20c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7f40157-96", "ovs_interfaceid": "c7f40157-961b-4250-a4b7-05b534b3d69b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.442039] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: f0b6273f-4177-44a3-8dd1-f65faeb07539] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.671868] env[61867]: DEBUG nova.scheduler.client.report [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1093.749711] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277197, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06624} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.749842] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1093.750727] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b4d96c2-8561-445e-bd87-025b2a67eba1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.772889] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 691b1a16-8cc4-4704-9041-ef3f6f5475e2/691b1a16-8cc4-4704-9041-ef3f6f5475e2.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1093.773228] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-973a1410-70dd-45ca-868c-35378e3a51ea {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.795030] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1093.795030] env[61867]: value = "task-1277198" [ 1093.795030] env[61867]: _type = "Task" [ 1093.795030] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.801060] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277198, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.897559] env[61867]: DEBUG oslo_concurrency.lockutils [req-9dc5cdc8-1fcf-44cb-af23-b451952aa0c4 req-3d89d486-0966-493e-b4f1-0772ad213a9f service nova] Releasing lock "refresh_cache-0b54931d-d5cd-459f-a6a2-e941c4f360d6" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.945444] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: c7203cea-2c2d-4a97-8ff4-3d33cd8443f7] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.178387] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.191s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.178908] env[61867]: DEBUG nova.compute.manager [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1094.303199] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277198, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.453860] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 81ada1e9-b100-44fd-8119-7ef8695e601f] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.689854] env[61867]: DEBUG nova.compute.utils [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1094.691277] env[61867]: DEBUG nova.compute.manager [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1094.691550] env[61867]: DEBUG nova.network.neutron [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1094.742176] env[61867]: DEBUG nova.policy [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1fc8ff85bced44c3a9de8b6c71d02bd5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '250e5d2fa5cb4a8a8bd0a6fdf9e50a3d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 1094.804623] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277198, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.960746] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: b5a6ff7c-2b48-4f82-ba53-ec1977736c8f] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.100768] env[61867]: DEBUG nova.network.neutron [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Successfully created port: 5c4e4ae5-fbad-4237-ae18-41bbf1446999 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1095.198731] env[61867]: DEBUG nova.compute.manager [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1095.304722] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277198, 'name': ReconfigVM_Task, 'duration_secs': 1.073902} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.305549] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 691b1a16-8cc4-4704-9041-ef3f6f5475e2/691b1a16-8cc4-4704-9041-ef3f6f5475e2.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1095.306416] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-41d113ae-8d6d-4e87-98b5-54cff31a4d5b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.319722] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1095.319722] env[61867]: value = "task-1277199" [ 1095.319722] env[61867]: _type = "Task" [ 1095.319722] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.330500] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277199, 'name': Rename_Task} progress is 10%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.468594] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: d0db58a5-d773-4e45-a238-07cfc156bdd2] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.834441] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277199, 'name': Rename_Task, 'duration_secs': 0.25516} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.834522] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1095.834792] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7f100f1c-8ebd-4233-877c-7f7c5247d564 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.842172] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1095.842172] env[61867]: value = "task-1277200" [ 1095.842172] env[61867]: _type = "Task" [ 1095.842172] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.848751] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277200, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.972151] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 7c0badee-0e2b-4a13-89b1-e57df5c2c9ef] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.215572] env[61867]: DEBUG nova.compute.manager [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1096.242143] env[61867]: DEBUG nova.virt.hardware [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1096.242407] env[61867]: DEBUG nova.virt.hardware [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1096.242571] env[61867]: DEBUG nova.virt.hardware [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1096.242757] env[61867]: DEBUG nova.virt.hardware [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1096.242905] env[61867]: DEBUG nova.virt.hardware [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1096.243068] env[61867]: DEBUG nova.virt.hardware [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1096.243285] env[61867]: DEBUG nova.virt.hardware [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1096.243445] env[61867]: DEBUG nova.virt.hardware [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1096.243675] env[61867]: DEBUG nova.virt.hardware [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1096.243844] env[61867]: DEBUG nova.virt.hardware [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1096.244028] env[61867]: DEBUG nova.virt.hardware [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1096.244896] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468f2316-3561-4bd5-a889-79579613ef79 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.253081] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b2437b-3b0c-42e8-a205-a11969c5e606 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.350879] env[61867]: DEBUG oslo_vmware.api [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277200, 'name': PowerOnVM_Task, 'duration_secs': 0.491403} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.351248] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1096.351549] env[61867]: INFO nova.compute.manager [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Took 9.56 seconds to spawn the instance on the hypervisor. [ 1096.351748] env[61867]: DEBUG nova.compute.manager [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1096.352542] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d790fefd-99e8-481c-9348-a98ad5dd1563 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.476148] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 8656c3b8-b7b1-442a-9719-1588ee83d19b] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.534677] env[61867]: DEBUG nova.compute.manager [req-a29f83b1-04bb-42a4-9c38-9e6261993cdd req-53ee3db6-3486-4a8e-9536-4ab591b09571 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Received event network-vif-plugged-5c4e4ae5-fbad-4237-ae18-41bbf1446999 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1096.535035] env[61867]: DEBUG oslo_concurrency.lockutils [req-a29f83b1-04bb-42a4-9c38-9e6261993cdd req-53ee3db6-3486-4a8e-9536-4ab591b09571 service nova] Acquiring lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.535257] env[61867]: DEBUG oslo_concurrency.lockutils [req-a29f83b1-04bb-42a4-9c38-9e6261993cdd req-53ee3db6-3486-4a8e-9536-4ab591b09571 service nova] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.537016] env[61867]: DEBUG oslo_concurrency.lockutils [req-a29f83b1-04bb-42a4-9c38-9e6261993cdd req-53ee3db6-3486-4a8e-9536-4ab591b09571 service nova] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.537016] env[61867]: DEBUG nova.compute.manager [req-a29f83b1-04bb-42a4-9c38-9e6261993cdd req-53ee3db6-3486-4a8e-9536-4ab591b09571 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] No waiting events found dispatching network-vif-plugged-5c4e4ae5-fbad-4237-ae18-41bbf1446999 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1096.537016] env[61867]: WARNING nova.compute.manager [req-a29f83b1-04bb-42a4-9c38-9e6261993cdd req-53ee3db6-3486-4a8e-9536-4ab591b09571 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Received unexpected event network-vif-plugged-5c4e4ae5-fbad-4237-ae18-41bbf1446999 for instance with vm_state building and task_state spawning. [ 1096.587942] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Acquiring lock "c47e3944-c182-4e55-b507-29caa6f4a26f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.587942] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Lock "c47e3944-c182-4e55-b507-29caa6f4a26f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.636649] env[61867]: DEBUG nova.network.neutron [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Successfully updated port: 5c4e4ae5-fbad-4237-ae18-41bbf1446999 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1096.869597] env[61867]: INFO nova.compute.manager [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Took 15.89 seconds to build instance. [ 1096.980108] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: b1ec32c5-0642-4dca-ad43-05a5172d04d7] Instance has had 0 of 5 cleanup attempts {{(pid=61867) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.093458] env[61867]: DEBUG nova.compute.manager [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Starting instance... {{(pid=61867) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1097.139138] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.139287] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.139373] env[61867]: DEBUG nova.network.neutron [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1097.147196] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "691b1a16-8cc4-4704-9041-ef3f6f5475e2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.373744] env[61867]: DEBUG oslo_concurrency.lockutils [None req-780e510a-d4b2-42b5-b4f8-e3e45282e7b0 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "691b1a16-8cc4-4704-9041-ef3f6f5475e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.394s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.373744] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "691b1a16-8cc4-4704-9041-ef3f6f5475e2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.226s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.373744] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "691b1a16-8cc4-4704-9041-ef3f6f5475e2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.373744] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "691b1a16-8cc4-4704-9041-ef3f6f5475e2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.373975] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "691b1a16-8cc4-4704-9041-ef3f6f5475e2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.376829] env[61867]: INFO nova.compute.manager [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Terminating instance [ 1097.378567] env[61867]: DEBUG nova.compute.manager [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1097.378567] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1097.379350] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-402583b9-0cfd-4535-962f-68911a159270 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.387033] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1097.387221] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba374fb4-fece-465f-a0c5-9910114fd9b1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.393425] env[61867]: DEBUG oslo_vmware.api [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1097.393425] env[61867]: value = "task-1277201" [ 1097.393425] env[61867]: _type = "Task" [ 1097.393425] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.401257] env[61867]: DEBUG oslo_vmware.api [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277201, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.483807] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.619570] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.619570] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.619570] env[61867]: INFO nova.compute.claims [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1097.679939] env[61867]: DEBUG nova.network.neutron [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1097.812997] env[61867]: DEBUG nova.network.neutron [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Updating instance_info_cache with network_info: [{"id": "5c4e4ae5-fbad-4237-ae18-41bbf1446999", "address": "fa:16:3e:21:31:aa", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c4e4ae5-fb", "ovs_interfaceid": "5c4e4ae5-fbad-4237-ae18-41bbf1446999", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.903434] env[61867]: DEBUG oslo_vmware.api [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277201, 'name': PowerOffVM_Task, 'duration_secs': 0.198923} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.903743] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1097.903933] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1097.904199] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-85bca092-8eda-4969-b97f-c2300b742650 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.964618] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1097.964885] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1097.965085] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Deleting the datastore file [datastore2] 691b1a16-8cc4-4704-9041-ef3f6f5475e2 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1097.965660] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ac0d5c7e-6f6e-495f-bc86-79e5ddbaa94e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.972770] env[61867]: DEBUG oslo_vmware.api [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1097.972770] env[61867]: value = "task-1277203" [ 1097.972770] env[61867]: _type = "Task" [ 1097.972770] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.980570] env[61867]: DEBUG oslo_vmware.api [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277203, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.238043] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1098.238141] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1098.315378] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.315643] env[61867]: DEBUG nova.compute.manager [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Instance network_info: |[{"id": "5c4e4ae5-fbad-4237-ae18-41bbf1446999", "address": "fa:16:3e:21:31:aa", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c4e4ae5-fb", "ovs_interfaceid": "5c4e4ae5-fbad-4237-ae18-41bbf1446999", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1098.316103] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:31:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '399f3826-705c-45f7-9fe0-3a08a945151a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5c4e4ae5-fbad-4237-ae18-41bbf1446999', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1098.323652] env[61867]: DEBUG oslo.service.loopingcall [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1098.324106] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1098.324375] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6e4d286a-2531-46a1-8503-5b25b6a4d13c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.344330] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1098.344330] env[61867]: value = "task-1277204" [ 1098.344330] env[61867]: _type = "Task" [ 1098.344330] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.351984] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277204, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.483086] env[61867]: DEBUG oslo_vmware.api [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277203, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156872} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.483397] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1098.483668] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1098.483903] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1098.484107] env[61867]: INFO nova.compute.manager [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1098.484355] env[61867]: DEBUG oslo.service.loopingcall [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1098.484581] env[61867]: DEBUG nova.compute.manager [-] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1098.484683] env[61867]: DEBUG nova.network.neutron [-] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1098.561680] env[61867]: DEBUG nova.compute.manager [req-09ea6956-5cf7-44f2-bb54-48b08d57396a req-5608ca09-cfe6-4659-8e78-fa46a157a3c0 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Received event network-changed-5c4e4ae5-fbad-4237-ae18-41bbf1446999 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1098.561999] env[61867]: DEBUG nova.compute.manager [req-09ea6956-5cf7-44f2-bb54-48b08d57396a req-5608ca09-cfe6-4659-8e78-fa46a157a3c0 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Refreshing instance network info cache due to event network-changed-5c4e4ae5-fbad-4237-ae18-41bbf1446999. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1098.562342] env[61867]: DEBUG oslo_concurrency.lockutils [req-09ea6956-5cf7-44f2-bb54-48b08d57396a req-5608ca09-cfe6-4659-8e78-fa46a157a3c0 service nova] Acquiring lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.562581] env[61867]: DEBUG oslo_concurrency.lockutils [req-09ea6956-5cf7-44f2-bb54-48b08d57396a req-5608ca09-cfe6-4659-8e78-fa46a157a3c0 service nova] Acquired lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.562839] env[61867]: DEBUG nova.network.neutron [req-09ea6956-5cf7-44f2-bb54-48b08d57396a req-5608ca09-cfe6-4659-8e78-fa46a157a3c0 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Refreshing network info cache for port 5c4e4ae5-fbad-4237-ae18-41bbf1446999 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1098.723501] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7622ccb4-3348-42ba-8988-89de37ce20ae {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.731140] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbabe7d2-ad9e-4476-8071-8f2d3129469b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.763229] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c070f517-0f13-46f8-b7c3-58ffc5ddf1fb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.765472] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1098.766062] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1098.770613] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f239ade-542a-4694-8d5d-a4d8a78f2a76 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.784895] env[61867]: DEBUG nova.compute.provider_tree [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.854124] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277204, 'name': CreateVM_Task, 'duration_secs': 0.439347} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.854305] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1098.855038] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.855221] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.855541] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1098.855788] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81e5918b-ed55-449f-8677-553a41d07fa2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.860116] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1098.860116] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526e51c8-cbe6-8fcd-5dc1-4fa6fa9f2715" [ 1098.860116] env[61867]: _type = "Task" [ 1098.860116] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.867536] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526e51c8-cbe6-8fcd-5dc1-4fa6fa9f2715, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.217638] env[61867]: DEBUG nova.network.neutron [-] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.269861] env[61867]: DEBUG nova.network.neutron [req-09ea6956-5cf7-44f2-bb54-48b08d57396a req-5608ca09-cfe6-4659-8e78-fa46a157a3c0 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Updated VIF entry in instance network info cache for port 5c4e4ae5-fbad-4237-ae18-41bbf1446999. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1099.270250] env[61867]: DEBUG nova.network.neutron [req-09ea6956-5cf7-44f2-bb54-48b08d57396a req-5608ca09-cfe6-4659-8e78-fa46a157a3c0 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Updating instance_info_cache with network_info: [{"id": "5c4e4ae5-fbad-4237-ae18-41bbf1446999", "address": "fa:16:3e:21:31:aa", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c4e4ae5-fb", "ovs_interfaceid": "5c4e4ae5-fbad-4237-ae18-41bbf1446999", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.287528] env[61867]: DEBUG nova.scheduler.client.report [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1099.370744] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]526e51c8-cbe6-8fcd-5dc1-4fa6fa9f2715, 'name': SearchDatastore_Task, 'duration_secs': 0.046691} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.371022] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1099.371260] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1099.371511] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1099.371644] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.371826] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1099.372129] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1f0d6db9-e744-4f14-800d-3f2b2ee627d2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.380151] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1099.380338] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1099.381024] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-caf1497c-94aa-4141-a61a-145658aca3e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.385919] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1099.385919] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520f5fe0-e42f-fa08-ba79-5cc1cd5e64da" [ 1099.385919] env[61867]: _type = "Task" [ 1099.385919] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.393118] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520f5fe0-e42f-fa08-ba79-5cc1cd5e64da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.721022] env[61867]: INFO nova.compute.manager [-] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Took 1.24 seconds to deallocate network for instance. [ 1099.772272] env[61867]: DEBUG oslo_concurrency.lockutils [req-09ea6956-5cf7-44f2-bb54-48b08d57396a req-5608ca09-cfe6-4659-8e78-fa46a157a3c0 service nova] Releasing lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1099.792566] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.175s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.793077] env[61867]: DEBUG nova.compute.manager [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Start building networks asynchronously for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1099.896414] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]520f5fe0-e42f-fa08-ba79-5cc1cd5e64da, 'name': SearchDatastore_Task, 'duration_secs': 0.007888} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.897172] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bab27e05-fd67-479c-8977-0c027c8326d2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.901908] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1099.901908] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d416ef-5277-dc48-26a9-f58088a013f3" [ 1099.901908] env[61867]: _type = "Task" [ 1099.901908] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.908835] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d416ef-5277-dc48-26a9-f58088a013f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.227708] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.228016] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.228244] env[61867]: DEBUG nova.objects.instance [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lazy-loading 'resources' on Instance uuid 691b1a16-8cc4-4704-9041-ef3f6f5475e2 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1100.297556] env[61867]: DEBUG nova.compute.utils [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1100.298934] env[61867]: DEBUG nova.compute.manager [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Allocating IP information in the background. {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1100.299116] env[61867]: DEBUG nova.network.neutron [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] allocate_for_instance() {{(pid=61867) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1100.340412] env[61867]: DEBUG nova.policy [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '957ce4b2cf1c49829bf685866d44da63', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '71684ead49ab4efcba275a1cca64cdeb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61867) authorize /opt/stack/nova/nova/policy.py:201}} [ 1100.414317] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d416ef-5277-dc48-26a9-f58088a013f3, 'name': SearchDatastore_Task, 'duration_secs': 0.011392} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.414598] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.414860] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66/214bb1c3-3d79-4ad8-9f1a-aad95dc83d66.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1100.415143] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4822ec48-33ad-46ce-9e65-7720485b1bef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.421152] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1100.421152] env[61867]: value = "task-1277205" [ 1100.421152] env[61867]: _type = "Task" [ 1100.421152] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.428521] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277205, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.589109] env[61867]: DEBUG nova.network.neutron [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Successfully created port: e5bc423d-603b-4c80-9b7d-539e13d21988 {{(pid=61867) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1100.594847] env[61867]: DEBUG nova.compute.manager [req-0a41420c-2597-4f40-9776-50939d6b2479 req-ffbdfe67-3d51-4373-8b6d-96a27fddcb33 service nova] [instance: 691b1a16-8cc4-4704-9041-ef3f6f5475e2] Received event network-vif-deleted-218c1590-49c5-460e-b6bf-03377d5a3c75 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1100.803462] env[61867]: DEBUG nova.compute.manager [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Start building block device mappings for instance. {{(pid=61867) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1100.823800] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d388e389-e4e1-48f9-b5ef-b07393a866c1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.835099] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f7bc158-7f4d-4511-82dc-fa3bcba9b176 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.865721] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82620911-2fe3-4ec6-ab9a-8092a35e5b91 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.873186] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-558bcc91-afd2-4a87-8a70-15ca2d34d6eb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.887091] env[61867]: DEBUG nova.compute.provider_tree [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.930168] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277205, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.435476} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.930445] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66/214bb1c3-3d79-4ad8-9f1a-aad95dc83d66.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1100.930664] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1100.930907] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc787c53-9272-416b-a47c-a3a40562c0e8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.937196] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1100.937196] env[61867]: value = "task-1277206" [ 1100.937196] env[61867]: _type = "Task" [ 1100.937196] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.946408] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277206, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.391155] env[61867]: DEBUG nova.scheduler.client.report [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1101.447508] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277206, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058034} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.447732] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1101.448509] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-062b3127-692c-43ec-9705-43e4cd835c33 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.469531] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66/214bb1c3-3d79-4ad8-9f1a-aad95dc83d66.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1101.469801] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0996f9ef-5e2b-4cb6-8254-d7e2030dab05 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.490205] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1101.490205] env[61867]: value = "task-1277207" [ 1101.490205] env[61867]: _type = "Task" [ 1101.490205] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.497691] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277207, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.814979] env[61867]: DEBUG nova.compute.manager [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Start spawning the instance on the hypervisor. {{(pid=61867) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1101.838791] env[61867]: DEBUG nova.virt.hardware [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T12:41:49Z,direct_url=,disk_format='vmdk',id=4ca02567-c128-482f-b204-6ac166ac3160,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84601e35c8e8487cb78fc16a2536a4c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T12:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1101.839063] env[61867]: DEBUG nova.virt.hardware [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1101.839246] env[61867]: DEBUG nova.virt.hardware [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1101.839439] env[61867]: DEBUG nova.virt.hardware [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1101.839590] env[61867]: DEBUG nova.virt.hardware [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1101.839742] env[61867]: DEBUG nova.virt.hardware [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1101.839951] env[61867]: DEBUG nova.virt.hardware [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1101.840141] env[61867]: DEBUG nova.virt.hardware [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1101.840321] env[61867]: DEBUG nova.virt.hardware [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1101.840488] env[61867]: DEBUG nova.virt.hardware [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1101.840663] env[61867]: DEBUG nova.virt.hardware [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1101.841523] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4712bba-54d0-4c40-823b-8768f945c76f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.849061] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c108a2e-8b3a-4aeb-99c9-a36d21e0aec2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.895310] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.667s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.916194] env[61867]: INFO nova.scheduler.client.report [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Deleted allocations for instance 691b1a16-8cc4-4704-9041-ef3f6f5475e2 [ 1101.999833] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277207, 'name': ReconfigVM_Task, 'duration_secs': 0.283493} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.999981] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66/214bb1c3-3d79-4ad8-9f1a-aad95dc83d66.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1102.001758] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-790f8f26-fe28-44d8-b4a0-cb55565ecd8e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.006959] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1102.006959] env[61867]: value = "task-1277208" [ 1102.006959] env[61867]: _type = "Task" [ 1102.006959] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.014709] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277208, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.089244] env[61867]: DEBUG nova.network.neutron [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Successfully updated port: e5bc423d-603b-4c80-9b7d-539e13d21988 {{(pid=61867) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1102.425965] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02970c3b-9eb0-4a0d-af37-d19d20855fc7 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "691b1a16-8cc4-4704-9041-ef3f6f5475e2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.053s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.517682] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277208, 'name': Rename_Task, 'duration_secs': 0.156733} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.517968] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1102.518224] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-98da1aea-3b5b-4304-9a96-9a10b258538d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.524900] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1102.524900] env[61867]: value = "task-1277209" [ 1102.524900] env[61867]: _type = "Task" [ 1102.524900] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.533013] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277209, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.593668] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Acquiring lock "refresh_cache-c47e3944-c182-4e55-b507-29caa6f4a26f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.593884] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Acquired lock "refresh_cache-c47e3944-c182-4e55-b507-29caa6f4a26f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.594067] env[61867]: DEBUG nova.network.neutron [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1102.624986] env[61867]: DEBUG nova.compute.manager [req-57646ee6-9242-4480-9941-f7015047a654 req-9d0b67a0-e6e1-464c-9def-746f664a0bcc service nova] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Received event network-vif-plugged-e5bc423d-603b-4c80-9b7d-539e13d21988 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1102.625236] env[61867]: DEBUG oslo_concurrency.lockutils [req-57646ee6-9242-4480-9941-f7015047a654 req-9d0b67a0-e6e1-464c-9def-746f664a0bcc service nova] Acquiring lock "c47e3944-c182-4e55-b507-29caa6f4a26f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.625446] env[61867]: DEBUG oslo_concurrency.lockutils [req-57646ee6-9242-4480-9941-f7015047a654 req-9d0b67a0-e6e1-464c-9def-746f664a0bcc service nova] Lock "c47e3944-c182-4e55-b507-29caa6f4a26f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.625613] env[61867]: DEBUG oslo_concurrency.lockutils [req-57646ee6-9242-4480-9941-f7015047a654 req-9d0b67a0-e6e1-464c-9def-746f664a0bcc service nova] Lock "c47e3944-c182-4e55-b507-29caa6f4a26f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.625783] env[61867]: DEBUG nova.compute.manager [req-57646ee6-9242-4480-9941-f7015047a654 req-9d0b67a0-e6e1-464c-9def-746f664a0bcc service nova] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] No waiting events found dispatching network-vif-plugged-e5bc423d-603b-4c80-9b7d-539e13d21988 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1102.625944] env[61867]: WARNING nova.compute.manager [req-57646ee6-9242-4480-9941-f7015047a654 req-9d0b67a0-e6e1-464c-9def-746f664a0bcc service nova] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Received unexpected event network-vif-plugged-e5bc423d-603b-4c80-9b7d-539e13d21988 for instance with vm_state building and task_state spawning. [ 1102.626116] env[61867]: DEBUG nova.compute.manager [req-57646ee6-9242-4480-9941-f7015047a654 req-9d0b67a0-e6e1-464c-9def-746f664a0bcc service nova] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Received event network-changed-e5bc423d-603b-4c80-9b7d-539e13d21988 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1102.626273] env[61867]: DEBUG nova.compute.manager [req-57646ee6-9242-4480-9941-f7015047a654 req-9d0b67a0-e6e1-464c-9def-746f664a0bcc service nova] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Refreshing instance network info cache due to event network-changed-e5bc423d-603b-4c80-9b7d-539e13d21988. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1102.626436] env[61867]: DEBUG oslo_concurrency.lockutils [req-57646ee6-9242-4480-9941-f7015047a654 req-9d0b67a0-e6e1-464c-9def-746f664a0bcc service nova] Acquiring lock "refresh_cache-c47e3944-c182-4e55-b507-29caa6f4a26f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.035128] env[61867]: DEBUG oslo_vmware.api [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277209, 'name': PowerOnVM_Task, 'duration_secs': 0.416502} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.035418] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1103.035701] env[61867]: INFO nova.compute.manager [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Took 6.82 seconds to spawn the instance on the hypervisor. [ 1103.035931] env[61867]: DEBUG nova.compute.manager [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1103.036819] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa40ba2-ec54-46ac-8ef4-adbf37e6437f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.125079] env[61867]: DEBUG nova.network.neutron [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1103.256731] env[61867]: DEBUG nova.network.neutron [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Updating instance_info_cache with network_info: [{"id": "e5bc423d-603b-4c80-9b7d-539e13d21988", "address": "fa:16:3e:5d:c8:76", "network": {"id": "8c1ae05f-8229-4198-b8da-346ac3073f24", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1423687714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71684ead49ab4efcba275a1cca64cdeb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5bc423d-60", "ovs_interfaceid": "e5bc423d-603b-4c80-9b7d-539e13d21988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.553934] env[61867]: INFO nova.compute.manager [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Took 11.59 seconds to build instance. [ 1103.759118] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Releasing lock "refresh_cache-c47e3944-c182-4e55-b507-29caa6f4a26f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.759482] env[61867]: DEBUG nova.compute.manager [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Instance network_info: |[{"id": "e5bc423d-603b-4c80-9b7d-539e13d21988", "address": "fa:16:3e:5d:c8:76", "network": {"id": "8c1ae05f-8229-4198-b8da-346ac3073f24", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1423687714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71684ead49ab4efcba275a1cca64cdeb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5bc423d-60", "ovs_interfaceid": "e5bc423d-603b-4c80-9b7d-539e13d21988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61867) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1103.759808] env[61867]: DEBUG oslo_concurrency.lockutils [req-57646ee6-9242-4480-9941-f7015047a654 req-9d0b67a0-e6e1-464c-9def-746f664a0bcc service nova] Acquired lock "refresh_cache-c47e3944-c182-4e55-b507-29caa6f4a26f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.759996] env[61867]: DEBUG nova.network.neutron [req-57646ee6-9242-4480-9941-f7015047a654 req-9d0b67a0-e6e1-464c-9def-746f664a0bcc service nova] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Refreshing network info cache for port e5bc423d-603b-4c80-9b7d-539e13d21988 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1103.761203] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5d:c8:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '40859343-2baa-45fd-88e3-ebf8aaed2b19', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e5bc423d-603b-4c80-9b7d-539e13d21988', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1103.769053] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Creating folder: Project (71684ead49ab4efcba275a1cca64cdeb). Parent ref: group-v274258. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1103.771489] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3550763f-08f3-4e55-960f-3d393016981c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.782209] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Created folder: Project (71684ead49ab4efcba275a1cca64cdeb) in parent group-v274258. [ 1103.782405] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Creating folder: Instances. Parent ref: group-v274433. {{(pid=61867) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1103.782636] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-74c50a0d-19f2-4869-a8f0-623b517a3de6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.790963] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Created folder: Instances in parent group-v274433. [ 1103.792026] env[61867]: DEBUG oslo.service.loopingcall [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1103.792026] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1103.792026] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ddf23c97-1da8-4b23-ae4f-c13a1600b00b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.811129] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1103.811129] env[61867]: value = "task-1277212" [ 1103.811129] env[61867]: _type = "Task" [ 1103.811129] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.818715] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277212, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.976030] env[61867]: DEBUG nova.network.neutron [req-57646ee6-9242-4480-9941-f7015047a654 req-9d0b67a0-e6e1-464c-9def-746f664a0bcc service nova] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Updated VIF entry in instance network info cache for port e5bc423d-603b-4c80-9b7d-539e13d21988. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1103.976448] env[61867]: DEBUG nova.network.neutron [req-57646ee6-9242-4480-9941-f7015047a654 req-9d0b67a0-e6e1-464c-9def-746f664a0bcc service nova] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Updating instance_info_cache with network_info: [{"id": "e5bc423d-603b-4c80-9b7d-539e13d21988", "address": "fa:16:3e:5d:c8:76", "network": {"id": "8c1ae05f-8229-4198-b8da-346ac3073f24", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1423687714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71684ead49ab4efcba275a1cca64cdeb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5bc423d-60", "ovs_interfaceid": "e5bc423d-603b-4c80-9b7d-539e13d21988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.056209] env[61867]: DEBUG oslo_concurrency.lockutils [None req-c37a8fc7-473f-44dc-beb2-37b893f570c4 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.096s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.278632] env[61867]: DEBUG nova.compute.manager [req-8dcc9345-3c42-407a-aa88-e9644b2f0ec4 req-f2311600-9d33-4860-bd0a-17ce4f33caa8 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Received event network-changed-5c4e4ae5-fbad-4237-ae18-41bbf1446999 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1104.278934] env[61867]: DEBUG nova.compute.manager [req-8dcc9345-3c42-407a-aa88-e9644b2f0ec4 req-f2311600-9d33-4860-bd0a-17ce4f33caa8 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Refreshing instance network info cache due to event network-changed-5c4e4ae5-fbad-4237-ae18-41bbf1446999. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1104.279097] env[61867]: DEBUG oslo_concurrency.lockutils [req-8dcc9345-3c42-407a-aa88-e9644b2f0ec4 req-f2311600-9d33-4860-bd0a-17ce4f33caa8 service nova] Acquiring lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.279248] env[61867]: DEBUG oslo_concurrency.lockutils [req-8dcc9345-3c42-407a-aa88-e9644b2f0ec4 req-f2311600-9d33-4860-bd0a-17ce4f33caa8 service nova] Acquired lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.279430] env[61867]: DEBUG nova.network.neutron [req-8dcc9345-3c42-407a-aa88-e9644b2f0ec4 req-f2311600-9d33-4860-bd0a-17ce4f33caa8 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Refreshing network info cache for port 5c4e4ae5-fbad-4237-ae18-41bbf1446999 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1104.321700] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277212, 'name': CreateVM_Task, 'duration_secs': 0.30226} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.321871] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1104.322541] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.322714] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.323082] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1104.323337] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4b5ffdc-7a3b-49f5-b69f-e9def72ce9b2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.327707] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Waiting for the task: (returnval){ [ 1104.327707] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d2562f-331e-37fa-867e-8bfcfff37885" [ 1104.327707] env[61867]: _type = "Task" [ 1104.327707] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.336025] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d2562f-331e-37fa-867e-8bfcfff37885, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.479509] env[61867]: DEBUG oslo_concurrency.lockutils [req-57646ee6-9242-4480-9941-f7015047a654 req-9d0b67a0-e6e1-464c-9def-746f664a0bcc service nova] Releasing lock "refresh_cache-c47e3944-c182-4e55-b507-29caa6f4a26f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.837731] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d2562f-331e-37fa-867e-8bfcfff37885, 'name': SearchDatastore_Task, 'duration_secs': 0.008883} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.838025] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.838247] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Processing image 4ca02567-c128-482f-b204-6ac166ac3160 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1104.838478] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.838626] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.838984] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1104.839087] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f98b8880-837a-42f2-9a66-d51fce60a9e2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.846934] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1104.847129] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1104.847803] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-407dd394-a314-49e1-87a2-3c5a88fcfd38 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.854608] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Waiting for the task: (returnval){ [ 1104.854608] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525b2fc2-580b-d3ca-33eb-930f4276a0d9" [ 1104.854608] env[61867]: _type = "Task" [ 1104.854608] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.861881] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525b2fc2-580b-d3ca-33eb-930f4276a0d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.986187] env[61867]: DEBUG nova.network.neutron [req-8dcc9345-3c42-407a-aa88-e9644b2f0ec4 req-f2311600-9d33-4860-bd0a-17ce4f33caa8 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Updated VIF entry in instance network info cache for port 5c4e4ae5-fbad-4237-ae18-41bbf1446999. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1104.986570] env[61867]: DEBUG nova.network.neutron [req-8dcc9345-3c42-407a-aa88-e9644b2f0ec4 req-f2311600-9d33-4860-bd0a-17ce4f33caa8 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Updating instance_info_cache with network_info: [{"id": "5c4e4ae5-fbad-4237-ae18-41bbf1446999", "address": "fa:16:3e:21:31:aa", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c4e4ae5-fb", "ovs_interfaceid": "5c4e4ae5-fbad-4237-ae18-41bbf1446999", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.364521] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]525b2fc2-580b-d3ca-33eb-930f4276a0d9, 'name': SearchDatastore_Task, 'duration_secs': 0.009138} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.365346] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1cbf4be-0a33-4fa2-91b3-3c0c541bb6e3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.370180] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Waiting for the task: (returnval){ [ 1105.370180] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529aace7-4a78-fea3-ddc0-0a72cb986873" [ 1105.370180] env[61867]: _type = "Task" [ 1105.370180] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.377238] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529aace7-4a78-fea3-ddc0-0a72cb986873, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.489116] env[61867]: DEBUG oslo_concurrency.lockutils [req-8dcc9345-3c42-407a-aa88-e9644b2f0ec4 req-f2311600-9d33-4860-bd0a-17ce4f33caa8 service nova] Releasing lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.880633] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]529aace7-4a78-fea3-ddc0-0a72cb986873, 'name': SearchDatastore_Task, 'duration_secs': 0.009003} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.880943] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.881202] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] c47e3944-c182-4e55-b507-29caa6f4a26f/c47e3944-c182-4e55-b507-29caa6f4a26f.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1105.881463] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9d15494d-80c1-4a09-801f-620985673529 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.887713] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Waiting for the task: (returnval){ [ 1105.887713] env[61867]: value = "task-1277213" [ 1105.887713] env[61867]: _type = "Task" [ 1105.887713] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.895811] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': task-1277213, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.398074] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': task-1277213, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.412712} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.398376] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4ca02567-c128-482f-b204-6ac166ac3160/4ca02567-c128-482f-b204-6ac166ac3160.vmdk to [datastore1] c47e3944-c182-4e55-b507-29caa6f4a26f/c47e3944-c182-4e55-b507-29caa6f4a26f.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1106.398599] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Extending root virtual disk to 1048576 {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1106.398860] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de7d8003-f5b5-448b-bc08-05c06705a9cc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.406125] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Waiting for the task: (returnval){ [ 1106.406125] env[61867]: value = "task-1277214" [ 1106.406125] env[61867]: _type = "Task" [ 1106.406125] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.413962] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': task-1277214, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.916368] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': task-1277214, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071549} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.916368] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Extended root virtual disk {{(pid=61867) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1106.916747] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f4b5ef0-a888-41a1-acbc-ce3b2303b552 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.938941] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] c47e3944-c182-4e55-b507-29caa6f4a26f/c47e3944-c182-4e55-b507-29caa6f4a26f.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1106.939242] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e3f37ff-bec4-48f3-bfd0-767a2eaa85d8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.963026] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Waiting for the task: (returnval){ [ 1106.963026] env[61867]: value = "task-1277215" [ 1106.963026] env[61867]: _type = "Task" [ 1106.963026] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.968496] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': task-1277215, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.472757] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': task-1277215, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.972025] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': task-1277215, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.473469] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': task-1277215, 'name': ReconfigVM_Task, 'duration_secs': 1.095518} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.473821] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Reconfigured VM instance instance-0000006f to attach disk [datastore1] c47e3944-c182-4e55-b507-29caa6f4a26f/c47e3944-c182-4e55-b507-29caa6f4a26f.vmdk or device None with type sparse {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1108.474467] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9382ceff-2b37-46ad-b20d-85ef4d94a896 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.482661] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Waiting for the task: (returnval){ [ 1108.482661] env[61867]: value = "task-1277216" [ 1108.482661] env[61867]: _type = "Task" [ 1108.482661] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.490313] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': task-1277216, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.992132] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': task-1277216, 'name': Rename_Task, 'duration_secs': 0.13438} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.992420] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1108.992671] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-be6c4cb3-ad1e-4866-bca1-a62b2a23fef3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.999264] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Waiting for the task: (returnval){ [ 1108.999264] env[61867]: value = "task-1277217" [ 1108.999264] env[61867]: _type = "Task" [ 1108.999264] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.006597] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': task-1277217, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.511499] env[61867]: DEBUG oslo_vmware.api [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': task-1277217, 'name': PowerOnVM_Task, 'duration_secs': 0.417945} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.511788] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1109.511995] env[61867]: INFO nova.compute.manager [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Took 7.70 seconds to spawn the instance on the hypervisor. [ 1109.512193] env[61867]: DEBUG nova.compute.manager [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1109.512946] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75ee7ce0-3603-4288-a22e-3459b28e10a0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.031952] env[61867]: INFO nova.compute.manager [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Took 12.43 seconds to build instance. [ 1110.534897] env[61867]: DEBUG oslo_concurrency.lockutils [None req-6c4aa7d0-79e9-4a4e-99df-08ee7204ac2e tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Lock "c47e3944-c182-4e55-b507-29caa6f4a26f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.948s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.306134] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Acquiring lock "c47e3944-c182-4e55-b507-29caa6f4a26f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.306439] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Lock "c47e3944-c182-4e55-b507-29caa6f4a26f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.306636] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Acquiring lock "c47e3944-c182-4e55-b507-29caa6f4a26f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.306864] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Lock "c47e3944-c182-4e55-b507-29caa6f4a26f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.307064] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Lock "c47e3944-c182-4e55-b507-29caa6f4a26f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.309554] env[61867]: INFO nova.compute.manager [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Terminating instance [ 1111.311415] env[61867]: DEBUG nova.compute.manager [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1111.311618] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1111.312446] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693b3624-d320-4224-bd4a-0928ae42c9ab {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.319819] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1111.320048] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8f2ebad-c670-44db-b9d8-619a5517305c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.326328] env[61867]: DEBUG oslo_vmware.api [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Waiting for the task: (returnval){ [ 1111.326328] env[61867]: value = "task-1277218" [ 1111.326328] env[61867]: _type = "Task" [ 1111.326328] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.333575] env[61867]: DEBUG oslo_vmware.api [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': task-1277218, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.837248] env[61867]: DEBUG oslo_vmware.api [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': task-1277218, 'name': PowerOffVM_Task, 'duration_secs': 0.15309} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.837531] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1111.837741] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1111.837999] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-60326b22-7137-491d-bb46-5af4b9a477f2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.895394] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1111.895716] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1111.895951] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Deleting the datastore file [datastore1] c47e3944-c182-4e55-b507-29caa6f4a26f {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1111.896267] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f3440b7c-9f7a-432d-b847-6345a4fa1ca7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.902658] env[61867]: DEBUG oslo_vmware.api [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Waiting for the task: (returnval){ [ 1111.902658] env[61867]: value = "task-1277220" [ 1111.902658] env[61867]: _type = "Task" [ 1111.902658] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.910136] env[61867]: DEBUG oslo_vmware.api [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': task-1277220, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.411453] env[61867]: DEBUG oslo_vmware.api [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Task: {'id': task-1277220, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156382} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.411824] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1112.412037] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1112.412226] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1112.412406] env[61867]: INFO nova.compute.manager [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1112.412648] env[61867]: DEBUG oslo.service.loopingcall [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1112.412842] env[61867]: DEBUG nova.compute.manager [-] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1112.412935] env[61867]: DEBUG nova.network.neutron [-] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1112.705534] env[61867]: DEBUG nova.compute.manager [req-6584ad9b-749a-47b9-b1ed-6c932a271205 req-6431188f-b06d-4e2a-8541-2e0bb86c2355 service nova] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Received event network-vif-deleted-e5bc423d-603b-4c80-9b7d-539e13d21988 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1112.705889] env[61867]: INFO nova.compute.manager [req-6584ad9b-749a-47b9-b1ed-6c932a271205 req-6431188f-b06d-4e2a-8541-2e0bb86c2355 service nova] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Neutron deleted interface e5bc423d-603b-4c80-9b7d-539e13d21988; detaching it from the instance and deleting it from the info cache [ 1112.706168] env[61867]: DEBUG nova.network.neutron [req-6584ad9b-749a-47b9-b1ed-6c932a271205 req-6431188f-b06d-4e2a-8541-2e0bb86c2355 service nova] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.176836] env[61867]: DEBUG nova.network.neutron [-] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.208587] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-30809976-5e6d-4bb5-b6c5-4771d65d2464 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.217735] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85857a8f-cf2b-44bc-9d99-4ac6b41c465d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.241214] env[61867]: DEBUG nova.compute.manager [req-6584ad9b-749a-47b9-b1ed-6c932a271205 req-6431188f-b06d-4e2a-8541-2e0bb86c2355 service nova] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Detach interface failed, port_id=e5bc423d-603b-4c80-9b7d-539e13d21988, reason: Instance c47e3944-c182-4e55-b507-29caa6f4a26f could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1113.434363] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "6fa6c260-9c20-4808-9dd7-946a27651681" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.434633] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "6fa6c260-9c20-4808-9dd7-946a27651681" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.434823] env[61867]: INFO nova.compute.manager [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Shelving [ 1113.679912] env[61867]: INFO nova.compute.manager [-] [instance: c47e3944-c182-4e55-b507-29caa6f4a26f] Took 1.27 seconds to deallocate network for instance. [ 1113.942881] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1113.943159] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-919f1fde-c858-415c-b527-d790ec6534b6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.950236] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1113.950236] env[61867]: value = "task-1277221" [ 1113.950236] env[61867]: _type = "Task" [ 1113.950236] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.958363] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277221, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.186736] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.187033] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.187265] env[61867]: DEBUG nova.objects.instance [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Lazy-loading 'resources' on Instance uuid c47e3944-c182-4e55-b507-29caa6f4a26f {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1114.460367] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277221, 'name': PowerOffVM_Task, 'duration_secs': 0.170682} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.460658] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1114.461391] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d03e6c-5913-4a31-833b-a1aa25c9da56 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.479792] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d035b762-f6ba-4ff6-989b-c0b3a2809d12 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.748091] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ff8b75-549f-465c-b7af-ff171752589e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.755099] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffbf1b1a-f2ac-4556-a2fc-103981d69d06 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.783703] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618d7c88-c131-4baf-b355-22c55f6431ce {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.790271] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea0d6248-3aef-4cc1-81d2-f33379257718 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.802770] env[61867]: DEBUG nova.compute.provider_tree [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.989770] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Creating Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1114.990068] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-75ef7edf-1c65-4f7e-9947-44d2a1da06c8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.998250] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1114.998250] env[61867]: value = "task-1277222" [ 1114.998250] env[61867]: _type = "Task" [ 1114.998250] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.008213] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277222, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.306295] env[61867]: DEBUG nova.scheduler.client.report [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1116.152088] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.965s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.159767] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277222, 'name': CreateSnapshot_Task, 'duration_secs': 0.431948} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.160054] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Created Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1116.160815] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf88e40-d133-40b8-aa3b-1ab919e34040 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.171841] env[61867]: INFO nova.scheduler.client.report [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Deleted allocations for instance c47e3944-c182-4e55-b507-29caa6f4a26f [ 1116.678080] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Creating linked-clone VM from snapshot {{(pid=61867) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1116.680339] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-df5c4fc9-962e-4ae6-a909-528a273d966b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.683522] env[61867]: DEBUG oslo_concurrency.lockutils [None req-cd98fd8d-754e-48a4-80c8-94f0fc1ea862 tempest-ServersNegativeTestMultiTenantJSON-942653276 tempest-ServersNegativeTestMultiTenantJSON-942653276-project-member] Lock "c47e3944-c182-4e55-b507-29caa6f4a26f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.377s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.689775] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1116.689775] env[61867]: value = "task-1277223" [ 1116.689775] env[61867]: _type = "Task" [ 1116.689775] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.697777] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277223, 'name': CloneVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.200154] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277223, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.700399] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277223, 'name': CloneVM_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.201122] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277223, 'name': CloneVM_Task, 'duration_secs': 1.03243} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.201441] env[61867]: INFO nova.virt.vmwareapi.vmops [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Created linked-clone VM from snapshot [ 1118.202205] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4794fc8-64b6-4c5d-be3d-20507d5e3534 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.209273] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Uploading image c653b476-1fca-4555-88c5-77f7a4fc53c2 {{(pid=61867) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1118.234532] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1118.234532] env[61867]: value = "vm-274437" [ 1118.234532] env[61867]: _type = "VirtualMachine" [ 1118.234532] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1118.234958] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b638b6c8-b487-4d8a-9621-fb2e427351e6 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.242070] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lease: (returnval){ [ 1118.242070] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52257a7c-2033-69cf-4ebd-73a463a9de70" [ 1118.242070] env[61867]: _type = "HttpNfcLease" [ 1118.242070] env[61867]: } obtained for exporting VM: (result){ [ 1118.242070] env[61867]: value = "vm-274437" [ 1118.242070] env[61867]: _type = "VirtualMachine" [ 1118.242070] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1118.242359] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the lease: (returnval){ [ 1118.242359] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52257a7c-2033-69cf-4ebd-73a463a9de70" [ 1118.242359] env[61867]: _type = "HttpNfcLease" [ 1118.242359] env[61867]: } to be ready. {{(pid=61867) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1118.248489] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1118.248489] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52257a7c-2033-69cf-4ebd-73a463a9de70" [ 1118.248489] env[61867]: _type = "HttpNfcLease" [ 1118.248489] env[61867]: } is initializing. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1118.749929] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1118.749929] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52257a7c-2033-69cf-4ebd-73a463a9de70" [ 1118.749929] env[61867]: _type = "HttpNfcLease" [ 1118.749929] env[61867]: } is ready. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1118.750904] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1118.750904] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52257a7c-2033-69cf-4ebd-73a463a9de70" [ 1118.750904] env[61867]: _type = "HttpNfcLease" [ 1118.750904] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1118.751042] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bacf748c-3865-4074-8bfa-c12e4a2b54f5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.758356] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db90fe-7847-8668-db8d-f9c0eb781d75/disk-0.vmdk from lease info. {{(pid=61867) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1118.758534] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db90fe-7847-8668-db8d-f9c0eb781d75/disk-0.vmdk for reading. {{(pid=61867) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1118.845317] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f0bbc0b6-800e-4370-a9f6-c1ba2f1cf148 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.973136] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1119.481114] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Getting list of instances from cluster (obj){ [ 1119.481114] env[61867]: value = "domain-c8" [ 1119.481114] env[61867]: _type = "ClusterComputeResource" [ 1119.481114] env[61867]: } {{(pid=61867) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1119.482646] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c46431-dce8-4e2b-bf80-ed65713b38fc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.496734] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Got total of 3 instances {{(pid=61867) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1119.496994] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Triggering sync for uuid 6fa6c260-9c20-4808-9dd7-946a27651681 {{(pid=61867) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1119.497264] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Triggering sync for uuid 0b54931d-d5cd-459f-a6a2-e941c4f360d6 {{(pid=61867) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1119.497468] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Triggering sync for uuid 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66 {{(pid=61867) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1119.497860] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "6fa6c260-9c20-4808-9dd7-946a27651681" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.498154] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.498411] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.498713] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.498960] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.499866] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b60bc0-2866-4711-8224-d447ef5ac396 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.503434] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3401c031-65a9-443d-8eb1-473b7abfb505 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.012529] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.514s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.015137] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.516s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.483965] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.484493] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.957248] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.957543] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.957929] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Starting heal instance info cache {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1125.964093] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Didn't find any instances for network info cache update. {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1125.964472] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1126.053218] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db90fe-7847-8668-db8d-f9c0eb781d75/disk-0.vmdk. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1126.054193] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc11099-d624-4536-9a2d-0bf604beb510 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.060170] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db90fe-7847-8668-db8d-f9c0eb781d75/disk-0.vmdk is in state: ready. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1126.060341] env[61867]: ERROR oslo_vmware.rw_handles [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db90fe-7847-8668-db8d-f9c0eb781d75/disk-0.vmdk due to incomplete transfer. [ 1126.060553] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c3b5003e-20e3-4a28-9629-11356ab32baa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.066977] env[61867]: DEBUG oslo_vmware.rw_handles [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db90fe-7847-8668-db8d-f9c0eb781d75/disk-0.vmdk. {{(pid=61867) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1126.067188] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Uploaded image c653b476-1fca-4555-88c5-77f7a4fc53c2 to the Glance image server {{(pid=61867) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1126.069337] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Destroying the VM {{(pid=61867) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1126.069566] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-31cec728-0156-4f63-ba5a-5fe86d0eff4e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.074823] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1126.074823] env[61867]: value = "task-1277225" [ 1126.074823] env[61867]: _type = "Task" [ 1126.074823] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.081885] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277225, 'name': Destroy_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.467771] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.468028] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.468204] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.468364] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61867) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1126.469288] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a52a83-458c-4421-b615-9a52e6374b14 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.477504] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5918338c-d8c9-4d05-a6c7-c67b0ce6fd11 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.491156] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a0ae93e-d79a-4b80-acc5-5f454605c058 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.498050] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129f8a14-0a8c-4abe-8a23-922cd41ab550 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.527734] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181199MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61867) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1126.527887] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.528194] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.585032] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277225, 'name': Destroy_Task, 'duration_secs': 0.367315} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.585333] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Destroyed the VM [ 1126.585583] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Deleting Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1126.585842] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-99df070c-b46e-4613-9709-4cbda932abe4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.592422] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1126.592422] env[61867]: value = "task-1277226" [ 1126.592422] env[61867]: _type = "Task" [ 1126.592422] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.600218] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277226, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.101490] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277226, 'name': RemoveSnapshot_Task, 'duration_secs': 0.363241} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.101853] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Deleted Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1127.102047] env[61867]: DEBUG nova.compute.manager [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1127.102776] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba2e8e8-1ff8-4f2a-a74e-eec830216c58 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.553763] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 6fa6c260-9c20-4808-9dd7-946a27651681 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1127.553979] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 0b54931d-d5cd-459f-a6a2-e941c4f360d6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1127.554133] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1127.554318] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1127.554457] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1127.603128] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d70ffbe-9272-454c-a265-f2fa908881b8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.610682] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d8f1d0-febc-4c66-95e1-13b66394ff12 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.616641] env[61867]: INFO nova.compute.manager [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Shelve offloading [ 1127.641910] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1127.642338] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d70d488-5c7b-4ea0-922e-70741ae32e65 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.644348] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-546fea0d-f59d-457f-892a-105d9c606592 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.652063] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29471c26-dd3b-43ac-9b0a-49edf6f28911 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.656742] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1127.656742] env[61867]: value = "task-1277227" [ 1127.656742] env[61867]: _type = "Task" [ 1127.656742] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.667230] env[61867]: DEBUG nova.compute.provider_tree [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1127.674046] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] VM already powered off {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1127.674242] env[61867]: DEBUG nova.compute.manager [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1127.675410] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23e66938-a1f8-49b7-92c2-c9c1198973db {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.680987] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.681176] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquired lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.681349] env[61867]: DEBUG nova.network.neutron [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1128.170151] env[61867]: DEBUG nova.scheduler.client.report [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1128.411374] env[61867]: DEBUG nova.network.neutron [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Updating instance_info_cache with network_info: [{"id": "5bc71c96-38d5-4cd8-92d4-46e8713e8389", "address": "fa:16:3e:55:cd:c4", "network": {"id": "8db0c704-eb45-4fca-8624-d21877ab5ab3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-74001547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f245b07249fa4e37ac97cf66dbf6678f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bc71c96-38", "ovs_interfaceid": "5bc71c96-38d5-4cd8-92d4-46e8713e8389", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.675039] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61867) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1128.675039] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.147s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.914539] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Releasing lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.165447] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1129.166376] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f2f211-cbbd-4b8e-a838-ad92b566af25 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.171620] env[61867]: DEBUG nova.compute.manager [req-47d2147c-e4b3-43b4-a35a-0344b51d1930 req-6e154fdc-3cc4-4dbf-8ee8-771d9c6dab1b service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Received event network-vif-unplugged-5bc71c96-38d5-4cd8-92d4-46e8713e8389 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1129.171891] env[61867]: DEBUG oslo_concurrency.lockutils [req-47d2147c-e4b3-43b4-a35a-0344b51d1930 req-6e154fdc-3cc4-4dbf-8ee8-771d9c6dab1b service nova] Acquiring lock "6fa6c260-9c20-4808-9dd7-946a27651681-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.172138] env[61867]: DEBUG oslo_concurrency.lockutils [req-47d2147c-e4b3-43b4-a35a-0344b51d1930 req-6e154fdc-3cc4-4dbf-8ee8-771d9c6dab1b service nova] Lock "6fa6c260-9c20-4808-9dd7-946a27651681-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.172498] env[61867]: DEBUG oslo_concurrency.lockutils [req-47d2147c-e4b3-43b4-a35a-0344b51d1930 req-6e154fdc-3cc4-4dbf-8ee8-771d9c6dab1b service nova] Lock "6fa6c260-9c20-4808-9dd7-946a27651681-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.172712] env[61867]: DEBUG nova.compute.manager [req-47d2147c-e4b3-43b4-a35a-0344b51d1930 req-6e154fdc-3cc4-4dbf-8ee8-771d9c6dab1b service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] No waiting events found dispatching network-vif-unplugged-5bc71c96-38d5-4cd8-92d4-46e8713e8389 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1129.172899] env[61867]: WARNING nova.compute.manager [req-47d2147c-e4b3-43b4-a35a-0344b51d1930 req-6e154fdc-3cc4-4dbf-8ee8-771d9c6dab1b service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Received unexpected event network-vif-unplugged-5bc71c96-38d5-4cd8-92d4-46e8713e8389 for instance with vm_state shelved and task_state shelving_offloading. [ 1129.178131] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1129.178368] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-01fe15ba-682c-4e4c-b8d1-e8166fdcecd7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.239710] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1129.239926] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1129.240112] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Deleting the datastore file [datastore2] 6fa6c260-9c20-4808-9dd7-946a27651681 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1129.240347] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4dfdca67-883f-4f8e-9e67-7a6767bc381d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.246353] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1129.246353] env[61867]: value = "task-1277229" [ 1129.246353] env[61867]: _type = "Task" [ 1129.246353] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.253653] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277229, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.583268] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.583513] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.668036] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1129.668264] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1129.668442] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1129.668577] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61867) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1129.755499] env[61867]: DEBUG oslo_vmware.api [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277229, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161014} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.755693] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1129.755882] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1129.756073] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1129.779577] env[61867]: INFO nova.scheduler.client.report [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Deleted allocations for instance 6fa6c260-9c20-4808-9dd7-946a27651681 [ 1130.086612] env[61867]: DEBUG nova.compute.utils [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Using /dev/sd instead of None {{(pid=61867) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1130.284584] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.284941] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.285132] env[61867]: DEBUG nova.objects.instance [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lazy-loading 'resources' on Instance uuid 6fa6c260-9c20-4808-9dd7-946a27651681 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1130.589883] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.788182] env[61867]: DEBUG nova.objects.instance [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lazy-loading 'numa_topology' on Instance uuid 6fa6c260-9c20-4808-9dd7-946a27651681 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1130.957235] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1131.191196] env[61867]: DEBUG nova.compute.manager [req-75e42a72-2dbb-4307-b1cd-ecff2f76176e req-4d0d2d5b-4e4b-4a81-b913-ffc034ea4711 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Received event network-changed-5bc71c96-38d5-4cd8-92d4-46e8713e8389 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1131.191196] env[61867]: DEBUG nova.compute.manager [req-75e42a72-2dbb-4307-b1cd-ecff2f76176e req-4d0d2d5b-4e4b-4a81-b913-ffc034ea4711 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Refreshing instance network info cache due to event network-changed-5bc71c96-38d5-4cd8-92d4-46e8713e8389. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1131.191267] env[61867]: DEBUG oslo_concurrency.lockutils [req-75e42a72-2dbb-4307-b1cd-ecff2f76176e req-4d0d2d5b-4e4b-4a81-b913-ffc034ea4711 service nova] Acquiring lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.191468] env[61867]: DEBUG oslo_concurrency.lockutils [req-75e42a72-2dbb-4307-b1cd-ecff2f76176e req-4d0d2d5b-4e4b-4a81-b913-ffc034ea4711 service nova] Acquired lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.191739] env[61867]: DEBUG nova.network.neutron [req-75e42a72-2dbb-4307-b1cd-ecff2f76176e req-4d0d2d5b-4e4b-4a81-b913-ffc034ea4711 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Refreshing network info cache for port 5bc71c96-38d5-4cd8-92d4-46e8713e8389 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1131.290805] env[61867]: DEBUG nova.objects.base [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Object Instance<6fa6c260-9c20-4808-9dd7-946a27651681> lazy-loaded attributes: resources,numa_topology {{(pid=61867) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1131.329079] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c63e6f5d-2677-47e5-a3a4-ec267336e760 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.337076] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea82ac63-0768-4560-b14f-bf24fbc4e409 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.366142] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-480295f8-419f-41f5-9572-d2869d436295 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.373277] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58e3e98f-37d8-41c3-890e-7847aae5a86d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.387520] env[61867]: DEBUG nova.compute.provider_tree [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1131.651655] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.651934] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.652222] env[61867]: INFO nova.compute.manager [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Attaching volume 53b0332e-7e1b-4233-974e-5d27501dd756 to /dev/sdb [ 1131.682841] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a87e46b-e6b5-4c87-ab11-42cc4e86817f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.689641] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd3e24e-0c99-4dab-92f6-b30895d7f140 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.703119] env[61867]: DEBUG nova.virt.block_device [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Updating existing volume attachment record: bc347da8-e4b2-43df-b44d-e6b173536c2b {{(pid=61867) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1131.890288] env[61867]: DEBUG nova.scheduler.client.report [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1131.897927] env[61867]: DEBUG nova.network.neutron [req-75e42a72-2dbb-4307-b1cd-ecff2f76176e req-4d0d2d5b-4e4b-4a81-b913-ffc034ea4711 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Updated VIF entry in instance network info cache for port 5bc71c96-38d5-4cd8-92d4-46e8713e8389. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1131.898376] env[61867]: DEBUG nova.network.neutron [req-75e42a72-2dbb-4307-b1cd-ecff2f76176e req-4d0d2d5b-4e4b-4a81-b913-ffc034ea4711 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Updating instance_info_cache with network_info: [{"id": "5bc71c96-38d5-4cd8-92d4-46e8713e8389", "address": "fa:16:3e:55:cd:c4", "network": {"id": "8db0c704-eb45-4fca-8624-d21877ab5ab3", "bridge": null, "label": "tempest-ServersNegativeTestJSON-74001547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f245b07249fa4e37ac97cf66dbf6678f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap5bc71c96-38", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.347305] env[61867]: DEBUG oslo_concurrency.lockutils [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "6fa6c260-9c20-4808-9dd7-946a27651681" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.395456] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.110s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.400507] env[61867]: DEBUG oslo_concurrency.lockutils [req-75e42a72-2dbb-4307-b1cd-ecff2f76176e req-4d0d2d5b-4e4b-4a81-b913-ffc034ea4711 service nova] Releasing lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.902561] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0cb265bd-2d0d-4078-bc63-1dce4f27f6c4 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "6fa6c260-9c20-4808-9dd7-946a27651681" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.468s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.903357] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "6fa6c260-9c20-4808-9dd7-946a27651681" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 13.406s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.903554] env[61867]: INFO nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] During sync_power_state the instance has a pending task (shelving_image_uploading). Skip. [ 1132.903729] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "6fa6c260-9c20-4808-9dd7-946a27651681" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.904188] env[61867]: DEBUG oslo_concurrency.lockutils [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "6fa6c260-9c20-4808-9dd7-946a27651681" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.557s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.904359] env[61867]: INFO nova.compute.manager [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Unshelving [ 1133.927908] env[61867]: DEBUG oslo_concurrency.lockutils [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.928161] env[61867]: DEBUG oslo_concurrency.lockutils [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.928402] env[61867]: DEBUG nova.objects.instance [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lazy-loading 'pci_requests' on Instance uuid 6fa6c260-9c20-4808-9dd7-946a27651681 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1134.433877] env[61867]: DEBUG nova.objects.instance [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lazy-loading 'numa_topology' on Instance uuid 6fa6c260-9c20-4808-9dd7-946a27651681 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1134.936366] env[61867]: INFO nova.compute.claims [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1135.992041] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92ba8d05-1130-44eb-80c3-1123b96179e1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.999062] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ec5d46-5bb6-46c6-8bf1-a541810f51a2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.027808] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae046fc9-5757-443b-9a8b-3ada4350b774 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.034254] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af319944-6260-4366-ad66-53437635e260 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.046544] env[61867]: DEBUG nova.compute.provider_tree [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1136.250924] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Volume attach. Driver type: vmdk {{(pid=61867) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1136.251153] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274438', 'volume_id': '53b0332e-7e1b-4233-974e-5d27501dd756', 'name': 'volume-53b0332e-7e1b-4233-974e-5d27501dd756', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0b54931d-d5cd-459f-a6a2-e941c4f360d6', 'attached_at': '', 'detached_at': '', 'volume_id': '53b0332e-7e1b-4233-974e-5d27501dd756', 'serial': '53b0332e-7e1b-4233-974e-5d27501dd756'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1136.252085] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a000603e-26b7-4bfc-afee-b3b3325451bd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.268478] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60940aa5-b9e1-4543-a4b4-6c03db71f991 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.293118] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] volume-53b0332e-7e1b-4233-974e-5d27501dd756/volume-53b0332e-7e1b-4233-974e-5d27501dd756.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1136.293346] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-481a53f1-97d8-4827-bcd3-aff164707b87 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.310625] env[61867]: DEBUG oslo_vmware.api [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1136.310625] env[61867]: value = "task-1277232" [ 1136.310625] env[61867]: _type = "Task" [ 1136.310625] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.318321] env[61867]: DEBUG oslo_vmware.api [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277232, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.549898] env[61867]: DEBUG nova.scheduler.client.report [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1136.820228] env[61867]: DEBUG oslo_vmware.api [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277232, 'name': ReconfigVM_Task, 'duration_secs': 0.29882} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.820397] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Reconfigured VM instance instance-0000006c to attach disk [datastore2] volume-53b0332e-7e1b-4233-974e-5d27501dd756/volume-53b0332e-7e1b-4233-974e-5d27501dd756.vmdk or device None with type thin {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1136.825371] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e82b5fc4-5887-4db0-8a14-237b9bd04e11 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.839638] env[61867]: DEBUG oslo_vmware.api [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1136.839638] env[61867]: value = "task-1277233" [ 1136.839638] env[61867]: _type = "Task" [ 1136.839638] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.846666] env[61867]: DEBUG oslo_vmware.api [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277233, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.054757] env[61867]: DEBUG oslo_concurrency.lockutils [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.126s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.083485] env[61867]: INFO nova.network.neutron [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Updating port 5bc71c96-38d5-4cd8-92d4-46e8713e8389 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1137.348684] env[61867]: DEBUG oslo_vmware.api [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277233, 'name': ReconfigVM_Task, 'duration_secs': 0.138719} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.348996] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274438', 'volume_id': '53b0332e-7e1b-4233-974e-5d27501dd756', 'name': 'volume-53b0332e-7e1b-4233-974e-5d27501dd756', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0b54931d-d5cd-459f-a6a2-e941c4f360d6', 'attached_at': '', 'detached_at': '', 'volume_id': '53b0332e-7e1b-4233-974e-5d27501dd756', 'serial': '53b0332e-7e1b-4233-974e-5d27501dd756'} {{(pid=61867) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1138.384011] env[61867]: DEBUG nova.objects.instance [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lazy-loading 'flavor' on Instance uuid 0b54931d-d5cd-459f-a6a2-e941c4f360d6 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1138.640279] env[61867]: DEBUG nova.compute.manager [req-783aed97-c9cb-4a8d-8982-f54efc1009f9 req-5449dc60-8046-4d29-9189-6c79145ed986 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Received event network-vif-plugged-5bc71c96-38d5-4cd8-92d4-46e8713e8389 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1138.640454] env[61867]: DEBUG oslo_concurrency.lockutils [req-783aed97-c9cb-4a8d-8982-f54efc1009f9 req-5449dc60-8046-4d29-9189-6c79145ed986 service nova] Acquiring lock "6fa6c260-9c20-4808-9dd7-946a27651681-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.640658] env[61867]: DEBUG oslo_concurrency.lockutils [req-783aed97-c9cb-4a8d-8982-f54efc1009f9 req-5449dc60-8046-4d29-9189-6c79145ed986 service nova] Lock "6fa6c260-9c20-4808-9dd7-946a27651681-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.640836] env[61867]: DEBUG oslo_concurrency.lockutils [req-783aed97-c9cb-4a8d-8982-f54efc1009f9 req-5449dc60-8046-4d29-9189-6c79145ed986 service nova] Lock "6fa6c260-9c20-4808-9dd7-946a27651681-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.641043] env[61867]: DEBUG nova.compute.manager [req-783aed97-c9cb-4a8d-8982-f54efc1009f9 req-5449dc60-8046-4d29-9189-6c79145ed986 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] No waiting events found dispatching network-vif-plugged-5bc71c96-38d5-4cd8-92d4-46e8713e8389 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1138.641241] env[61867]: WARNING nova.compute.manager [req-783aed97-c9cb-4a8d-8982-f54efc1009f9 req-5449dc60-8046-4d29-9189-6c79145ed986 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Received unexpected event network-vif-plugged-5bc71c96-38d5-4cd8-92d4-46e8713e8389 for instance with vm_state shelved_offloaded and task_state spawning. [ 1138.730285] env[61867]: DEBUG oslo_concurrency.lockutils [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1138.730477] env[61867]: DEBUG oslo_concurrency.lockutils [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquired lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.730658] env[61867]: DEBUG nova.network.neutron [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1138.889359] env[61867]: DEBUG oslo_concurrency.lockutils [None req-a3f55745-ab61-4039-97e9-8b4f269243b5 tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.237s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.075525] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.075815] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.430271] env[61867]: DEBUG nova.network.neutron [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Updating instance_info_cache with network_info: [{"id": "5bc71c96-38d5-4cd8-92d4-46e8713e8389", "address": "fa:16:3e:55:cd:c4", "network": {"id": "8db0c704-eb45-4fca-8624-d21877ab5ab3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-74001547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f245b07249fa4e37ac97cf66dbf6678f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bc71c96-38", "ovs_interfaceid": "5bc71c96-38d5-4cd8-92d4-46e8713e8389", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.578991] env[61867]: INFO nova.compute.manager [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Detaching volume 53b0332e-7e1b-4233-974e-5d27501dd756 [ 1139.608371] env[61867]: INFO nova.virt.block_device [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Attempting to driver detach volume 53b0332e-7e1b-4233-974e-5d27501dd756 from mountpoint /dev/sdb [ 1139.608627] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Volume detach. Driver type: vmdk {{(pid=61867) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1139.608820] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274438', 'volume_id': '53b0332e-7e1b-4233-974e-5d27501dd756', 'name': 'volume-53b0332e-7e1b-4233-974e-5d27501dd756', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0b54931d-d5cd-459f-a6a2-e941c4f360d6', 'attached_at': '', 'detached_at': '', 'volume_id': '53b0332e-7e1b-4233-974e-5d27501dd756', 'serial': '53b0332e-7e1b-4233-974e-5d27501dd756'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1139.609719] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a768919-1759-48f8-80f5-71e8b845a240 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.630525] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f50f884-67fd-4876-8329-6efb5e0051fd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.636795] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-517108ae-1363-4135-a299-353b1a8999e1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.655627] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cda44cb5-26cf-44ff-96c4-577032c2e387 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.669390] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] The volume has not been displaced from its original location: [datastore2] volume-53b0332e-7e1b-4233-974e-5d27501dd756/volume-53b0332e-7e1b-4233-974e-5d27501dd756.vmdk. No consolidation needed. {{(pid=61867) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1139.674528] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Reconfiguring VM instance instance-0000006c to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1139.674777] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aebc17c1-b0a1-4ab7-b473-be3ee6cc92b3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.693629] env[61867]: DEBUG oslo_vmware.api [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1139.693629] env[61867]: value = "task-1277234" [ 1139.693629] env[61867]: _type = "Task" [ 1139.693629] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.700956] env[61867]: DEBUG oslo_vmware.api [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277234, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.932815] env[61867]: DEBUG oslo_concurrency.lockutils [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Releasing lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1139.959240] env[61867]: DEBUG nova.virt.hardware [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='bd8fa53c4d242cbd876e78eccb268d68',container_format='bare',created_at=2024-10-12T12:53:51Z,direct_url=,disk_format='vmdk',id=c653b476-1fca-4555-88c5-77f7a4fc53c2,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-802417726-shelved',owner='f245b07249fa4e37ac97cf66dbf6678f',properties=ImageMetaProps,protected=,size=31660032,status='active',tags=,updated_at=2024-10-12T12:54:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1139.959509] env[61867]: DEBUG nova.virt.hardware [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1139.959670] env[61867]: DEBUG nova.virt.hardware [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1139.959859] env[61867]: DEBUG nova.virt.hardware [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1139.960025] env[61867]: DEBUG nova.virt.hardware [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1139.960189] env[61867]: DEBUG nova.virt.hardware [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1139.960407] env[61867]: DEBUG nova.virt.hardware [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1139.960573] env[61867]: DEBUG nova.virt.hardware [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1139.960743] env[61867]: DEBUG nova.virt.hardware [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1139.960911] env[61867]: DEBUG nova.virt.hardware [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1139.961110] env[61867]: DEBUG nova.virt.hardware [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1139.962254] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-928bbea3-23a5-4d0b-8b4b-3fb09aa69607 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.969881] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf65bda5-a9ee-4d19-bb09-fc99f32c2ba8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.983789] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:cd:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91c1da19-ab68-4127-bacd-accbaff19651', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5bc71c96-38d5-4cd8-92d4-46e8713e8389', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1139.990868] env[61867]: DEBUG oslo.service.loopingcall [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1139.991108] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1139.991313] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-64630267-7751-40df-a1d9-37e2cf1eeb21 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.008387] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1140.008387] env[61867]: value = "task-1277235" [ 1140.008387] env[61867]: _type = "Task" [ 1140.008387] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.015179] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277235, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.204120] env[61867]: DEBUG oslo_vmware.api [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277234, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.517753] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277235, 'name': CreateVM_Task, 'duration_secs': 0.280275} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.518154] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1140.518615] env[61867]: DEBUG oslo_concurrency.lockutils [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c653b476-1fca-4555-88c5-77f7a4fc53c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.518792] env[61867]: DEBUG oslo_concurrency.lockutils [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c653b476-1fca-4555-88c5-77f7a4fc53c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.519222] env[61867]: DEBUG oslo_concurrency.lockutils [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c653b476-1fca-4555-88c5-77f7a4fc53c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1140.519452] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10cecb7a-e001-43f1-baa5-480f9ffd1678 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.523742] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1140.523742] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d6746f-9eb4-9361-44cd-dbb1150e11df" [ 1140.523742] env[61867]: _type = "Task" [ 1140.523742] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.530894] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52d6746f-9eb4-9361-44cd-dbb1150e11df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.663757] env[61867]: DEBUG nova.compute.manager [req-03e5a34d-49b9-4a19-9a1e-771c964e5355 req-3862f1c3-1ff1-4cba-963a-2502838ca20e service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Received event network-changed-5bc71c96-38d5-4cd8-92d4-46e8713e8389 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1140.663892] env[61867]: DEBUG nova.compute.manager [req-03e5a34d-49b9-4a19-9a1e-771c964e5355 req-3862f1c3-1ff1-4cba-963a-2502838ca20e service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Refreshing instance network info cache due to event network-changed-5bc71c96-38d5-4cd8-92d4-46e8713e8389. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1140.664155] env[61867]: DEBUG oslo_concurrency.lockutils [req-03e5a34d-49b9-4a19-9a1e-771c964e5355 req-3862f1c3-1ff1-4cba-963a-2502838ca20e service nova] Acquiring lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.664316] env[61867]: DEBUG oslo_concurrency.lockutils [req-03e5a34d-49b9-4a19-9a1e-771c964e5355 req-3862f1c3-1ff1-4cba-963a-2502838ca20e service nova] Acquired lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.664481] env[61867]: DEBUG nova.network.neutron [req-03e5a34d-49b9-4a19-9a1e-771c964e5355 req-3862f1c3-1ff1-4cba-963a-2502838ca20e service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Refreshing network info cache for port 5bc71c96-38d5-4cd8-92d4-46e8713e8389 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1140.704559] env[61867]: DEBUG oslo_vmware.api [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277234, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.034078] env[61867]: DEBUG oslo_concurrency.lockutils [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c653b476-1fca-4555-88c5-77f7a4fc53c2" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.034334] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Processing image c653b476-1fca-4555-88c5-77f7a4fc53c2 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1141.034573] env[61867]: DEBUG oslo_concurrency.lockutils [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c653b476-1fca-4555-88c5-77f7a4fc53c2/c653b476-1fca-4555-88c5-77f7a4fc53c2.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.034727] env[61867]: DEBUG oslo_concurrency.lockutils [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c653b476-1fca-4555-88c5-77f7a4fc53c2/c653b476-1fca-4555-88c5-77f7a4fc53c2.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.034914] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1141.035212] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6c002ce3-96d1-49f3-8af0-0cf6dc37a41b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.042700] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1141.042880] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1141.043548] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abfcd291-3929-406c-8aa4-0503f2738e2d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.048109] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1141.048109] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52be67f2-0b9f-e8ff-1093-be53c7a666b5" [ 1141.048109] env[61867]: _type = "Task" [ 1141.048109] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.054772] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52be67f2-0b9f-e8ff-1093-be53c7a666b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.204838] env[61867]: DEBUG oslo_vmware.api [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277234, 'name': ReconfigVM_Task, 'duration_secs': 1.210774} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.205132] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Reconfigured VM instance instance-0000006c to detach disk 2001 {{(pid=61867) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1141.209819] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2066d717-374c-4a28-8a19-20caa02ed1d5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.224343] env[61867]: DEBUG oslo_vmware.api [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1141.224343] env[61867]: value = "task-1277236" [ 1141.224343] env[61867]: _type = "Task" [ 1141.224343] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.231878] env[61867]: DEBUG oslo_vmware.api [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277236, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.365352] env[61867]: DEBUG nova.network.neutron [req-03e5a34d-49b9-4a19-9a1e-771c964e5355 req-3862f1c3-1ff1-4cba-963a-2502838ca20e service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Updated VIF entry in instance network info cache for port 5bc71c96-38d5-4cd8-92d4-46e8713e8389. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1141.365751] env[61867]: DEBUG nova.network.neutron [req-03e5a34d-49b9-4a19-9a1e-771c964e5355 req-3862f1c3-1ff1-4cba-963a-2502838ca20e service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Updating instance_info_cache with network_info: [{"id": "5bc71c96-38d5-4cd8-92d4-46e8713e8389", "address": "fa:16:3e:55:cd:c4", "network": {"id": "8db0c704-eb45-4fca-8624-d21877ab5ab3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-74001547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f245b07249fa4e37ac97cf66dbf6678f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bc71c96-38", "ovs_interfaceid": "5bc71c96-38d5-4cd8-92d4-46e8713e8389", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.559183] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Preparing fetch location {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1141.559517] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Fetch image to [datastore2] OSTACK_IMG_4eb9aad7-7013-406c-82e6-38ad3e84beed/OSTACK_IMG_4eb9aad7-7013-406c-82e6-38ad3e84beed.vmdk {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1141.560485] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Downloading stream optimized image c653b476-1fca-4555-88c5-77f7a4fc53c2 to [datastore2] OSTACK_IMG_4eb9aad7-7013-406c-82e6-38ad3e84beed/OSTACK_IMG_4eb9aad7-7013-406c-82e6-38ad3e84beed.vmdk on the data store datastore2 as vApp {{(pid=61867) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1141.560485] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Downloading image file data c653b476-1fca-4555-88c5-77f7a4fc53c2 to the ESX as VM named 'OSTACK_IMG_4eb9aad7-7013-406c-82e6-38ad3e84beed' {{(pid=61867) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1141.623968] env[61867]: DEBUG oslo_vmware.rw_handles [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1141.623968] env[61867]: value = "resgroup-9" [ 1141.623968] env[61867]: _type = "ResourcePool" [ 1141.623968] env[61867]: }. {{(pid=61867) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1141.624323] env[61867]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-21979813-3b3e-4c96-9710-7c89a63aadfa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.644757] env[61867]: DEBUG oslo_vmware.rw_handles [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lease: (returnval){ [ 1141.644757] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e20293-9a23-7bbe-b8d2-a0092a737f9f" [ 1141.644757] env[61867]: _type = "HttpNfcLease" [ 1141.644757] env[61867]: } obtained for vApp import into resource pool (val){ [ 1141.644757] env[61867]: value = "resgroup-9" [ 1141.644757] env[61867]: _type = "ResourcePool" [ 1141.644757] env[61867]: }. {{(pid=61867) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1141.645149] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the lease: (returnval){ [ 1141.645149] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e20293-9a23-7bbe-b8d2-a0092a737f9f" [ 1141.645149] env[61867]: _type = "HttpNfcLease" [ 1141.645149] env[61867]: } to be ready. {{(pid=61867) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1141.650527] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1141.650527] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e20293-9a23-7bbe-b8d2-a0092a737f9f" [ 1141.650527] env[61867]: _type = "HttpNfcLease" [ 1141.650527] env[61867]: } is initializing. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1141.733971] env[61867]: DEBUG oslo_vmware.api [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277236, 'name': ReconfigVM_Task, 'duration_secs': 0.136123} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.734379] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274438', 'volume_id': '53b0332e-7e1b-4233-974e-5d27501dd756', 'name': 'volume-53b0332e-7e1b-4233-974e-5d27501dd756', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0b54931d-d5cd-459f-a6a2-e941c4f360d6', 'attached_at': '', 'detached_at': '', 'volume_id': '53b0332e-7e1b-4233-974e-5d27501dd756', 'serial': '53b0332e-7e1b-4233-974e-5d27501dd756'} {{(pid=61867) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1141.868301] env[61867]: DEBUG oslo_concurrency.lockutils [req-03e5a34d-49b9-4a19-9a1e-771c964e5355 req-3862f1c3-1ff1-4cba-963a-2502838ca20e service nova] Releasing lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.985468] env[61867]: DEBUG oslo_concurrency.lockutils [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.985760] env[61867]: DEBUG oslo_concurrency.lockutils [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.985956] env[61867]: INFO nova.compute.manager [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Shelving [ 1142.152537] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1142.152537] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e20293-9a23-7bbe-b8d2-a0092a737f9f" [ 1142.152537] env[61867]: _type = "HttpNfcLease" [ 1142.152537] env[61867]: } is ready. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1142.152831] env[61867]: DEBUG oslo_vmware.rw_handles [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1142.152831] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52e20293-9a23-7bbe-b8d2-a0092a737f9f" [ 1142.152831] env[61867]: _type = "HttpNfcLease" [ 1142.152831] env[61867]: }. {{(pid=61867) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1142.153505] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e98243b7-6053-427b-bf08-5221c4205424 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.160325] env[61867]: DEBUG oslo_vmware.rw_handles [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52396bea-8c11-c001-8be9-5bc381912bf2/disk-0.vmdk from lease info. {{(pid=61867) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1142.160508] env[61867]: DEBUG oslo_vmware.rw_handles [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Creating HTTP connection to write to file with size = 31660032 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52396bea-8c11-c001-8be9-5bc381912bf2/disk-0.vmdk. {{(pid=61867) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1142.223985] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d95be58d-49f4-43b4-bb05-a02e0c875987 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.277909] env[61867]: DEBUG nova.objects.instance [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lazy-loading 'flavor' on Instance uuid 0b54931d-d5cd-459f-a6a2-e941c4f360d6 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1142.494632] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1142.494904] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e9b3d09-27af-48a9-ab55-2b72fbc2e771 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.503324] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1142.503324] env[61867]: value = "task-1277238" [ 1142.503324] env[61867]: _type = "Task" [ 1142.503324] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.513754] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277238, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.015494] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277238, 'name': PowerOffVM_Task, 'duration_secs': 0.192052} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.017311] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1143.018134] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b702c8c-bf22-46eb-be0a-d11d065491f7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.043632] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94223e95-0936-46e1-a887-3ed863b5d82f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.250585] env[61867]: DEBUG oslo_vmware.rw_handles [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Completed reading data from the image iterator. {{(pid=61867) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1143.250900] env[61867]: DEBUG oslo_vmware.rw_handles [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52396bea-8c11-c001-8be9-5bc381912bf2/disk-0.vmdk. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1143.251998] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3849fe6b-089d-4ffc-a259-f76fcce746c8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.258827] env[61867]: DEBUG oslo_vmware.rw_handles [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52396bea-8c11-c001-8be9-5bc381912bf2/disk-0.vmdk is in state: ready. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1143.259075] env[61867]: DEBUG oslo_vmware.rw_handles [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52396bea-8c11-c001-8be9-5bc381912bf2/disk-0.vmdk. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1143.259334] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-ae9195b3-e7f0-466c-8a0d-38c90f82dff9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.285875] env[61867]: DEBUG oslo_concurrency.lockutils [None req-fa896876-fd15-46b2-a543-6b3fea60125b tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.210s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.444288] env[61867]: DEBUG oslo_vmware.rw_handles [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52396bea-8c11-c001-8be9-5bc381912bf2/disk-0.vmdk. {{(pid=61867) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1143.444554] env[61867]: INFO nova.virt.vmwareapi.images [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Downloaded image file data c653b476-1fca-4555-88c5-77f7a4fc53c2 [ 1143.445414] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aab5e74-835e-4d10-8873-6ed75ea2f066 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.462513] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4a7280e8-11c8-4899-be74-e5af8970fb06 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.484494] env[61867]: INFO nova.virt.vmwareapi.images [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] The imported VM was unregistered [ 1143.486845] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Caching image {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1143.487096] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Creating directory with path [datastore2] devstack-image-cache_base/c653b476-1fca-4555-88c5-77f7a4fc53c2 {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1143.487362] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d5452dc-b43d-4f97-b81f-5844d4950588 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.497656] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Created directory with path [datastore2] devstack-image-cache_base/c653b476-1fca-4555-88c5-77f7a4fc53c2 {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1143.497842] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_4eb9aad7-7013-406c-82e6-38ad3e84beed/OSTACK_IMG_4eb9aad7-7013-406c-82e6-38ad3e84beed.vmdk to [datastore2] devstack-image-cache_base/c653b476-1fca-4555-88c5-77f7a4fc53c2/c653b476-1fca-4555-88c5-77f7a4fc53c2.vmdk. {{(pid=61867) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1143.498098] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-680e6033-129c-4b4c-9576-0ca6df6cc20a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.504412] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1143.504412] env[61867]: value = "task-1277240" [ 1143.504412] env[61867]: _type = "Task" [ 1143.504412] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.511584] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277240, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.558455] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Creating Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1143.559145] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-05c29a24-f5a0-4e45-a3a9-0f2dedf51a19 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.565841] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1143.565841] env[61867]: value = "task-1277241" [ 1143.565841] env[61867]: _type = "Task" [ 1143.565841] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.574411] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277241, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.014468] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277240, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.075559] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277241, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.340688] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.340989] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.341290] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.341585] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.341865] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.344519] env[61867]: INFO nova.compute.manager [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Terminating instance [ 1144.346675] env[61867]: DEBUG nova.compute.manager [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1144.346930] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1144.347827] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2018d0f3-b7c2-4816-869c-0091988caf1c {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.356234] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1144.356528] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7f3bea52-e63d-4f27-9767-637d926810c2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.363173] env[61867]: DEBUG oslo_vmware.api [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1144.363173] env[61867]: value = "task-1277242" [ 1144.363173] env[61867]: _type = "Task" [ 1144.363173] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.372047] env[61867]: DEBUG oslo_vmware.api [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277242, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.515979] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277240, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.576871] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277241, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.874574] env[61867]: DEBUG oslo_vmware.api [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277242, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.016755] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277240, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.077748] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277241, 'name': CreateSnapshot_Task, 'duration_secs': 1.10724} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.078259] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Created Snapshot of the VM instance {{(pid=61867) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1145.078892] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a3857b-e166-49c4-9cac-cac9de62d739 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.374408] env[61867]: DEBUG oslo_vmware.api [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277242, 'name': PowerOffVM_Task, 'duration_secs': 0.960439} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.374711] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1145.374902] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1145.375216] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a6e0d9c7-08af-4b03-a706-ef34d1f6c95d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.519060] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277240, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.597718] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Creating linked-clone VM from snapshot {{(pid=61867) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1145.598153] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7f4dd0e7-e5f3-4ffa-822a-293ef22d488a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.608427] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1145.608427] env[61867]: value = "task-1277244" [ 1145.608427] env[61867]: _type = "Task" [ 1145.608427] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.619391] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277244, 'name': CloneVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.780353] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1145.780611] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1145.780793] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Deleting the datastore file [datastore2] 0b54931d-d5cd-459f-a6a2-e941c4f360d6 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1145.781243] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fcd79258-30bc-42b1-9cd5-70a5e85dd069 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.788455] env[61867]: DEBUG oslo_vmware.api [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for the task: (returnval){ [ 1145.788455] env[61867]: value = "task-1277245" [ 1145.788455] env[61867]: _type = "Task" [ 1145.788455] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.797374] env[61867]: DEBUG oslo_vmware.api [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277245, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.022654] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277240, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.118939] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277244, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.298953] env[61867]: DEBUG oslo_vmware.api [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277245, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.518544] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277240, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.618976] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277244, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.799479] env[61867]: DEBUG oslo_vmware.api [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277245, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.020225] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277240, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.368986} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.020225] env[61867]: INFO nova.virt.vmwareapi.ds_util [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_4eb9aad7-7013-406c-82e6-38ad3e84beed/OSTACK_IMG_4eb9aad7-7013-406c-82e6-38ad3e84beed.vmdk to [datastore2] devstack-image-cache_base/c653b476-1fca-4555-88c5-77f7a4fc53c2/c653b476-1fca-4555-88c5-77f7a4fc53c2.vmdk. [ 1147.020225] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Cleaning up location [datastore2] OSTACK_IMG_4eb9aad7-7013-406c-82e6-38ad3e84beed {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1147.020225] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_4eb9aad7-7013-406c-82e6-38ad3e84beed {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1147.020521] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6da77562-4344-47cc-9075-748e9a1cecba {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.026164] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1147.026164] env[61867]: value = "task-1277246" [ 1147.026164] env[61867]: _type = "Task" [ 1147.026164] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.033580] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277246, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.119923] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277244, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.301603] env[61867]: DEBUG oslo_vmware.api [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277245, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.536448] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277246, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.278561} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.536741] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1147.536917] env[61867]: DEBUG oslo_concurrency.lockutils [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c653b476-1fca-4555-88c5-77f7a4fc53c2/c653b476-1fca-4555-88c5-77f7a4fc53c2.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.537210] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c653b476-1fca-4555-88c5-77f7a4fc53c2/c653b476-1fca-4555-88c5-77f7a4fc53c2.vmdk to [datastore2] 6fa6c260-9c20-4808-9dd7-946a27651681/6fa6c260-9c20-4808-9dd7-946a27651681.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1147.537488] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bddbb598-80b3-4e0f-87df-2437ee40f6e0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.543595] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1147.543595] env[61867]: value = "task-1277247" [ 1147.543595] env[61867]: _type = "Task" [ 1147.543595] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.550791] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277247, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.621877] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277244, 'name': CloneVM_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.802174] env[61867]: DEBUG oslo_vmware.api [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Task: {'id': task-1277245, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.557974} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.802553] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1147.802772] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1147.802990] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1147.803208] env[61867]: INFO nova.compute.manager [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Took 3.46 seconds to destroy the instance on the hypervisor. [ 1147.803517] env[61867]: DEBUG oslo.service.loopingcall [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1147.803738] env[61867]: DEBUG nova.compute.manager [-] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1147.803834] env[61867]: DEBUG nova.network.neutron [-] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1148.053835] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277247, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.122878] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277244, 'name': CloneVM_Task} progress is 95%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.307498] env[61867]: DEBUG nova.compute.manager [req-909b7b53-56d1-4516-958e-028e02bca182 req-a896c768-3f6f-44b7-b54c-14fbcd808253 service nova] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Received event network-vif-deleted-c7f40157-961b-4250-a4b7-05b534b3d69b {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1148.307604] env[61867]: INFO nova.compute.manager [req-909b7b53-56d1-4516-958e-028e02bca182 req-a896c768-3f6f-44b7-b54c-14fbcd808253 service nova] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Neutron deleted interface c7f40157-961b-4250-a4b7-05b534b3d69b; detaching it from the instance and deleting it from the info cache [ 1148.307766] env[61867]: DEBUG nova.network.neutron [req-909b7b53-56d1-4516-958e-028e02bca182 req-a896c768-3f6f-44b7-b54c-14fbcd808253 service nova] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.555880] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277247, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.621807] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277244, 'name': CloneVM_Task, 'duration_secs': 2.951521} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.622257] env[61867]: INFO nova.virt.vmwareapi.vmops [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Created linked-clone VM from snapshot [ 1148.623055] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04118a4c-3df8-419a-a9a8-28b2d18668ff {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.630428] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Uploading image e53f4285-1fdc-4f6b-b607-bfb1722ced7f {{(pid=61867) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1148.656860] env[61867]: DEBUG oslo_vmware.rw_handles [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1148.656860] env[61867]: value = "vm-274442" [ 1148.656860] env[61867]: _type = "VirtualMachine" [ 1148.656860] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1148.657321] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e6a0f488-20ee-43b4-8c74-9f5d9f30261d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.664845] env[61867]: DEBUG oslo_vmware.rw_handles [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lease: (returnval){ [ 1148.664845] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528472ef-59fd-c419-778c-08dd5e6cf1bd" [ 1148.664845] env[61867]: _type = "HttpNfcLease" [ 1148.664845] env[61867]: } obtained for exporting VM: (result){ [ 1148.664845] env[61867]: value = "vm-274442" [ 1148.664845] env[61867]: _type = "VirtualMachine" [ 1148.664845] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1148.665157] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the lease: (returnval){ [ 1148.665157] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528472ef-59fd-c419-778c-08dd5e6cf1bd" [ 1148.665157] env[61867]: _type = "HttpNfcLease" [ 1148.665157] env[61867]: } to be ready. {{(pid=61867) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1148.671887] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1148.671887] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528472ef-59fd-c419-778c-08dd5e6cf1bd" [ 1148.671887] env[61867]: _type = "HttpNfcLease" [ 1148.671887] env[61867]: } is initializing. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1148.784703] env[61867]: DEBUG nova.network.neutron [-] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.810269] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a91bcd7d-8710-47d8-894c-f54268a09efc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.820378] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1584de-7271-43a2-81cb-57039902a340 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.845434] env[61867]: DEBUG nova.compute.manager [req-909b7b53-56d1-4516-958e-028e02bca182 req-a896c768-3f6f-44b7-b54c-14fbcd808253 service nova] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Detach interface failed, port_id=c7f40157-961b-4250-a4b7-05b534b3d69b, reason: Instance 0b54931d-d5cd-459f-a6a2-e941c4f360d6 could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1149.055412] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277247, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.173579] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1149.173579] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528472ef-59fd-c419-778c-08dd5e6cf1bd" [ 1149.173579] env[61867]: _type = "HttpNfcLease" [ 1149.173579] env[61867]: } is ready. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1149.173943] env[61867]: DEBUG oslo_vmware.rw_handles [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1149.173943] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]528472ef-59fd-c419-778c-08dd5e6cf1bd" [ 1149.173943] env[61867]: _type = "HttpNfcLease" [ 1149.173943] env[61867]: }. {{(pid=61867) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1149.174794] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7397d3c-2ff0-4355-874b-5fca6482ef6e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.183710] env[61867]: DEBUG oslo_vmware.rw_handles [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526acc97-4ccb-cb2f-a761-d9952fd84e6c/disk-0.vmdk from lease info. {{(pid=61867) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1149.183911] env[61867]: DEBUG oslo_vmware.rw_handles [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526acc97-4ccb-cb2f-a761-d9952fd84e6c/disk-0.vmdk for reading. {{(pid=61867) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1149.287843] env[61867]: INFO nova.compute.manager [-] [instance: 0b54931d-d5cd-459f-a6a2-e941c4f360d6] Took 1.48 seconds to deallocate network for instance. [ 1149.295493] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0ef20347-7fc2-4d4e-b2da-9420a3a375f0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.557117] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277247, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.798976] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.799344] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.799654] env[61867]: DEBUG nova.objects.instance [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lazy-loading 'resources' on Instance uuid 0b54931d-d5cd-459f-a6a2-e941c4f360d6 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1150.058068] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277247, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.246719} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.058486] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c653b476-1fca-4555-88c5-77f7a4fc53c2/c653b476-1fca-4555-88c5-77f7a4fc53c2.vmdk to [datastore2] 6fa6c260-9c20-4808-9dd7-946a27651681/6fa6c260-9c20-4808-9dd7-946a27651681.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1150.059391] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-520b69a5-307c-48bc-8128-06975e9ee9c2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.082969] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 6fa6c260-9c20-4808-9dd7-946a27651681/6fa6c260-9c20-4808-9dd7-946a27651681.vmdk or device None with type streamOptimized {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1150.083405] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dee610b1-1708-473e-a471-fe58fc7073a4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.103819] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1150.103819] env[61867]: value = "task-1277249" [ 1150.103819] env[61867]: _type = "Task" [ 1150.103819] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.112312] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277249, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.366063] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0939dd-aa9e-4d58-985c-ff551d5a2469 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.374268] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8dae23a-9743-43bb-a4b4-a67948ca148b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.407697] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d15034d-b605-4f67-94e4-b17cac20dd58 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.415205] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b14422-9afa-4412-9078-78d57f19b8a9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.428379] env[61867]: DEBUG nova.compute.provider_tree [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1150.613861] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277249, 'name': ReconfigVM_Task, 'duration_secs': 0.278702} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.615111] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 6fa6c260-9c20-4808-9dd7-946a27651681/6fa6c260-9c20-4808-9dd7-946a27651681.vmdk or device None with type streamOptimized {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1150.615288] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a85f9f9a-b294-4053-a447-8a48edbcb23d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.621576] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1150.621576] env[61867]: value = "task-1277250" [ 1150.621576] env[61867]: _type = "Task" [ 1150.621576] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.629369] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277250, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.931573] env[61867]: DEBUG nova.scheduler.client.report [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1151.132514] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277250, 'name': Rename_Task, 'duration_secs': 0.139918} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.132514] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1151.133265] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-00697761-4073-4dd4-a44a-f8697eb82473 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.139122] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1151.139122] env[61867]: value = "task-1277251" [ 1151.139122] env[61867]: _type = "Task" [ 1151.139122] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.146577] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277251, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.436712] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.637s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.459954] env[61867]: INFO nova.scheduler.client.report [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Deleted allocations for instance 0b54931d-d5cd-459f-a6a2-e941c4f360d6 [ 1151.648162] env[61867]: DEBUG oslo_vmware.api [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277251, 'name': PowerOnVM_Task, 'duration_secs': 0.442148} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.648443] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1151.809994] env[61867]: DEBUG nova.compute.manager [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1151.810989] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea42ad4d-8e22-4797-9f0b-2ab68492ca00 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.968580] env[61867]: DEBUG oslo_concurrency.lockutils [None req-caa22cfd-0568-4202-a749-7e2130af1e9a tempest-AttachVolumeNegativeTest-593760199 tempest-AttachVolumeNegativeTest-593760199-project-member] Lock "0b54931d-d5cd-459f-a6a2-e941c4f360d6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.627s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.329073] env[61867]: DEBUG oslo_concurrency.lockutils [None req-630ef771-1dd9-451d-8bd3-4260fd8b4b96 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "6fa6c260-9c20-4808-9dd7-946a27651681" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.425s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.820749] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64eae698-8a21-4783-99cf-8508e9ccc6ef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.827657] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-7ca837f5-a6f0-43b5-afca-580e8bb6328e tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Suspending the VM {{(pid=61867) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1153.827917] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-42f8277a-b772-44a7-b4d4-50a628cc1218 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.833943] env[61867]: DEBUG oslo_vmware.api [None req-7ca837f5-a6f0-43b5-afca-580e8bb6328e tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1153.833943] env[61867]: value = "task-1277253" [ 1153.833943] env[61867]: _type = "Task" [ 1153.833943] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.842274] env[61867]: DEBUG oslo_vmware.api [None req-7ca837f5-a6f0-43b5-afca-580e8bb6328e tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277253, 'name': SuspendVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.344938] env[61867]: DEBUG oslo_vmware.api [None req-7ca837f5-a6f0-43b5-afca-580e8bb6328e tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277253, 'name': SuspendVM_Task} progress is 75%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.844389] env[61867]: DEBUG oslo_vmware.api [None req-7ca837f5-a6f0-43b5-afca-580e8bb6328e tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277253, 'name': SuspendVM_Task, 'duration_secs': 0.570534} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.844804] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-7ca837f5-a6f0-43b5-afca-580e8bb6328e tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Suspended the VM {{(pid=61867) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1154.844846] env[61867]: DEBUG nova.compute.manager [None req-7ca837f5-a6f0-43b5-afca-580e8bb6328e tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1154.845645] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83beb74e-4ddb-4fbe-80d8-daf28942ac4d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.265262] env[61867]: INFO nova.compute.manager [None req-13e0a2b3-2e04-4468-b854-50d35875b194 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Resuming [ 1156.265904] env[61867]: DEBUG nova.objects.instance [None req-13e0a2b3-2e04-4468-b854-50d35875b194 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lazy-loading 'flavor' on Instance uuid 6fa6c260-9c20-4808-9dd7-946a27651681 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1156.655588] env[61867]: DEBUG oslo_vmware.rw_handles [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526acc97-4ccb-cb2f-a761-d9952fd84e6c/disk-0.vmdk. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1156.656523] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0687c3ec-5c28-4817-b6e8-9e8ebd1f313b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.662317] env[61867]: DEBUG oslo_vmware.rw_handles [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526acc97-4ccb-cb2f-a761-d9952fd84e6c/disk-0.vmdk is in state: ready. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1156.662483] env[61867]: ERROR oslo_vmware.rw_handles [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526acc97-4ccb-cb2f-a761-d9952fd84e6c/disk-0.vmdk due to incomplete transfer. [ 1156.662702] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6c1bbd18-cfad-4861-aeea-e87c60d6bdd7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.669145] env[61867]: DEBUG oslo_vmware.rw_handles [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526acc97-4ccb-cb2f-a761-d9952fd84e6c/disk-0.vmdk. {{(pid=61867) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1156.669344] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Uploaded image e53f4285-1fdc-4f6b-b607-bfb1722ced7f to the Glance image server {{(pid=61867) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1156.671494] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Destroying the VM {{(pid=61867) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1156.671718] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-24f8ccb6-2f78-4f50-ad37-eed021b5594d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.677230] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1156.677230] env[61867]: value = "task-1277255" [ 1156.677230] env[61867]: _type = "Task" [ 1156.677230] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.685066] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277255, 'name': Destroy_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.187122] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277255, 'name': Destroy_Task, 'duration_secs': 0.339263} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.187369] env[61867]: INFO nova.virt.vmwareapi.vm_util [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Destroyed the VM [ 1157.187610] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Deleting Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1157.187855] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-2d85555f-5ed7-4cf2-9828-b734061f0716 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.193976] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1157.193976] env[61867]: value = "task-1277256" [ 1157.193976] env[61867]: _type = "Task" [ 1157.193976] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.201083] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277256, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.274702] env[61867]: DEBUG oslo_concurrency.lockutils [None req-13e0a2b3-2e04-4468-b854-50d35875b194 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1157.275760] env[61867]: DEBUG oslo_concurrency.lockutils [None req-13e0a2b3-2e04-4468-b854-50d35875b194 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquired lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.275760] env[61867]: DEBUG nova.network.neutron [None req-13e0a2b3-2e04-4468-b854-50d35875b194 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1157.703683] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277256, 'name': RemoveSnapshot_Task, 'duration_secs': 0.335103} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.703983] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Deleted Snapshot of the VM instance {{(pid=61867) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1157.704343] env[61867]: DEBUG nova.compute.manager [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1157.705109] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9978eaa-a2d8-4a43-b90e-ce7a899f04d1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.966091] env[61867]: DEBUG nova.network.neutron [None req-13e0a2b3-2e04-4468-b854-50d35875b194 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Updating instance_info_cache with network_info: [{"id": "5bc71c96-38d5-4cd8-92d4-46e8713e8389", "address": "fa:16:3e:55:cd:c4", "network": {"id": "8db0c704-eb45-4fca-8624-d21877ab5ab3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-74001547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f245b07249fa4e37ac97cf66dbf6678f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bc71c96-38", "ovs_interfaceid": "5bc71c96-38d5-4cd8-92d4-46e8713e8389", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1158.216880] env[61867]: INFO nova.compute.manager [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Shelve offloading [ 1158.218452] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1158.218699] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-63ce12a5-dd17-4e1b-bbd8-565a6bd90794 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.226484] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1158.226484] env[61867]: value = "task-1277257" [ 1158.226484] env[61867]: _type = "Task" [ 1158.226484] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.234295] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277257, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.469482] env[61867]: DEBUG oslo_concurrency.lockutils [None req-13e0a2b3-2e04-4468-b854-50d35875b194 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Releasing lock "refresh_cache-6fa6c260-9c20-4808-9dd7-946a27651681" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.470498] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa256c2-8aef-4155-a426-981e57476f73 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.477142] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-13e0a2b3-2e04-4468-b854-50d35875b194 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Resuming the VM {{(pid=61867) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1158.477404] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a86793c-3a93-430b-8c87-8d15b493ed08 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.482793] env[61867]: DEBUG oslo_vmware.api [None req-13e0a2b3-2e04-4468-b854-50d35875b194 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1158.482793] env[61867]: value = "task-1277259" [ 1158.482793] env[61867]: _type = "Task" [ 1158.482793] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.490078] env[61867]: DEBUG oslo_vmware.api [None req-13e0a2b3-2e04-4468-b854-50d35875b194 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277259, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.737427] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] VM already powered off {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1158.737577] env[61867]: DEBUG nova.compute.manager [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1158.738336] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ebe6245-6296-4dcb-88ed-1691da433ed3 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.743644] env[61867]: DEBUG oslo_concurrency.lockutils [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.743811] env[61867]: DEBUG oslo_concurrency.lockutils [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.743982] env[61867]: DEBUG nova.network.neutron [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1158.998095] env[61867]: DEBUG oslo_vmware.api [None req-13e0a2b3-2e04-4468-b854-50d35875b194 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277259, 'name': PowerOnVM_Task, 'duration_secs': 0.511745} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.998095] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-13e0a2b3-2e04-4468-b854-50d35875b194 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Resumed the VM {{(pid=61867) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1158.998095] env[61867]: DEBUG nova.compute.manager [None req-13e0a2b3-2e04-4468-b854-50d35875b194 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1158.998998] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64f2298c-75ef-4abe-85ad-08306cb6c64d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.463381] env[61867]: DEBUG nova.network.neutron [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Updating instance_info_cache with network_info: [{"id": "5c4e4ae5-fbad-4237-ae18-41bbf1446999", "address": "fa:16:3e:21:31:aa", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c4e4ae5-fb", "ovs_interfaceid": "5c4e4ae5-fbad-4237-ae18-41bbf1446999", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.965882] env[61867]: DEBUG oslo_concurrency.lockutils [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1160.200843] env[61867]: DEBUG nova.compute.manager [req-60f0c4ba-a0b4-4ea2-afa0-302b38eb7f0c req-812e93c6-98c5-40a6-ba91-2136e1311324 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Received event network-vif-unplugged-5c4e4ae5-fbad-4237-ae18-41bbf1446999 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1160.201100] env[61867]: DEBUG oslo_concurrency.lockutils [req-60f0c4ba-a0b4-4ea2-afa0-302b38eb7f0c req-812e93c6-98c5-40a6-ba91-2136e1311324 service nova] Acquiring lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.201327] env[61867]: DEBUG oslo_concurrency.lockutils [req-60f0c4ba-a0b4-4ea2-afa0-302b38eb7f0c req-812e93c6-98c5-40a6-ba91-2136e1311324 service nova] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.201512] env[61867]: DEBUG oslo_concurrency.lockutils [req-60f0c4ba-a0b4-4ea2-afa0-302b38eb7f0c req-812e93c6-98c5-40a6-ba91-2136e1311324 service nova] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.201637] env[61867]: DEBUG nova.compute.manager [req-60f0c4ba-a0b4-4ea2-afa0-302b38eb7f0c req-812e93c6-98c5-40a6-ba91-2136e1311324 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] No waiting events found dispatching network-vif-unplugged-5c4e4ae5-fbad-4237-ae18-41bbf1446999 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1160.201816] env[61867]: WARNING nova.compute.manager [req-60f0c4ba-a0b4-4ea2-afa0-302b38eb7f0c req-812e93c6-98c5-40a6-ba91-2136e1311324 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Received unexpected event network-vif-unplugged-5c4e4ae5-fbad-4237-ae18-41bbf1446999 for instance with vm_state shelved and task_state shelving_offloading. [ 1160.288307] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1160.289264] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fd4db3a-07f0-4e08-92ad-75158997b496 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.298370] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1160.298680] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7afdb433-8b84-45e9-9b31-597484a1187e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.365726] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1160.365964] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Deleting contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1160.366166] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Deleting the datastore file [datastore1] 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1160.366530] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cad1cef4-a07d-4133-9221-784592df461d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.372165] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1160.372165] env[61867]: value = "task-1277261" [ 1160.372165] env[61867]: _type = "Task" [ 1160.372165] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.380296] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277261, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.881805] env[61867]: DEBUG oslo_vmware.api [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277261, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122269} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.881805] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1160.881805] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Deleted contents of the VM from datastore datastore1 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1160.881987] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1160.904660] env[61867]: INFO nova.scheduler.client.report [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Deleted allocations for instance 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66 [ 1161.410071] env[61867]: DEBUG oslo_concurrency.lockutils [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.410071] env[61867]: DEBUG oslo_concurrency.lockutils [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.410071] env[61867]: DEBUG nova.objects.instance [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lazy-loading 'resources' on Instance uuid 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1161.914047] env[61867]: DEBUG nova.objects.instance [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lazy-loading 'numa_topology' on Instance uuid 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1162.232546] env[61867]: DEBUG nova.compute.manager [req-4272e1a3-850a-4511-95f1-96d84d125287 req-253d1dff-3c83-4378-a109-834fbaced694 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Received event network-changed-5c4e4ae5-fbad-4237-ae18-41bbf1446999 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1162.232832] env[61867]: DEBUG nova.compute.manager [req-4272e1a3-850a-4511-95f1-96d84d125287 req-253d1dff-3c83-4378-a109-834fbaced694 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Refreshing instance network info cache due to event network-changed-5c4e4ae5-fbad-4237-ae18-41bbf1446999. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1162.233279] env[61867]: DEBUG oslo_concurrency.lockutils [req-4272e1a3-850a-4511-95f1-96d84d125287 req-253d1dff-3c83-4378-a109-834fbaced694 service nova] Acquiring lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1162.233508] env[61867]: DEBUG oslo_concurrency.lockutils [req-4272e1a3-850a-4511-95f1-96d84d125287 req-253d1dff-3c83-4378-a109-834fbaced694 service nova] Acquired lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.233755] env[61867]: DEBUG nova.network.neutron [req-4272e1a3-850a-4511-95f1-96d84d125287 req-253d1dff-3c83-4378-a109-834fbaced694 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Refreshing network info cache for port 5c4e4ae5-fbad-4237-ae18-41bbf1446999 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1162.415929] env[61867]: DEBUG nova.objects.base [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Object Instance<214bb1c3-3d79-4ad8-9f1a-aad95dc83d66> lazy-loaded attributes: resources,numa_topology {{(pid=61867) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1162.455704] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c4e6a1-f39e-4434-970e-75a3a9ca581b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.463498] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-034463f6-ccc4-47e5-8908-92868dc8de70 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.493850] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3338c2-246d-4aed-96f5-0460d30b64fc {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.501111] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ccf937-7e6e-4c2e-8453-8620782dbabd {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.514741] env[61867]: DEBUG nova.compute.provider_tree [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1163.018956] env[61867]: DEBUG nova.scheduler.client.report [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1163.111504] env[61867]: DEBUG nova.network.neutron [req-4272e1a3-850a-4511-95f1-96d84d125287 req-253d1dff-3c83-4378-a109-834fbaced694 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Updated VIF entry in instance network info cache for port 5c4e4ae5-fbad-4237-ae18-41bbf1446999. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1163.111870] env[61867]: DEBUG nova.network.neutron [req-4272e1a3-850a-4511-95f1-96d84d125287 req-253d1dff-3c83-4378-a109-834fbaced694 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Updating instance_info_cache with network_info: [{"id": "5c4e4ae5-fbad-4237-ae18-41bbf1446999", "address": "fa:16:3e:21:31:aa", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap5c4e4ae5-fb", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1163.524264] env[61867]: DEBUG oslo_concurrency.lockutils [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.115s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.531484] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.589215] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "6fa6c260-9c20-4808-9dd7-946a27651681" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.589583] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "6fa6c260-9c20-4808-9dd7-946a27651681" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.589876] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "6fa6c260-9c20-4808-9dd7-946a27651681-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.590741] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "6fa6c260-9c20-4808-9dd7-946a27651681-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.590741] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "6fa6c260-9c20-4808-9dd7-946a27651681-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.592529] env[61867]: INFO nova.compute.manager [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Terminating instance [ 1163.594379] env[61867]: DEBUG nova.compute.manager [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1163.594584] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1163.595529] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b65130-55f8-444a-a2e4-6efe8dffeef9 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.603227] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1163.603467] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f797c2ad-d32a-4533-87c8-2dd3fdc96d9a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.610520] env[61867]: DEBUG oslo_vmware.api [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1163.610520] env[61867]: value = "task-1277262" [ 1163.610520] env[61867]: _type = "Task" [ 1163.610520] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.614173] env[61867]: DEBUG oslo_concurrency.lockutils [req-4272e1a3-850a-4511-95f1-96d84d125287 req-253d1dff-3c83-4378-a109-834fbaced694 service nova] Releasing lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1163.619079] env[61867]: DEBUG oslo_vmware.api [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277262, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.033525] env[61867]: DEBUG oslo_concurrency.lockutils [None req-da370ae5-1080-4677-8aed-fb923c4a3726 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 22.048s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.034505] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.503s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.034737] env[61867]: INFO nova.compute.manager [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Unshelving [ 1164.120071] env[61867]: DEBUG oslo_vmware.api [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277262, 'name': PowerOffVM_Task, 'duration_secs': 0.207856} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.120350] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1164.120524] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1164.120772] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6fa3c9c7-fd77-4146-b8ef-124395c50377 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.183945] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1164.184197] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1164.184435] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Deleting the datastore file [datastore2] 6fa6c260-9c20-4808-9dd7-946a27651681 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1164.184704] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d29c987b-6323-41a3-bfe9-66c728d32a93 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.191068] env[61867]: DEBUG oslo_vmware.api [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for the task: (returnval){ [ 1164.191068] env[61867]: value = "task-1277264" [ 1164.191068] env[61867]: _type = "Task" [ 1164.191068] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.200345] env[61867]: DEBUG oslo_vmware.api [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277264, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.700685] env[61867]: DEBUG oslo_vmware.api [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Task: {'id': task-1277264, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126031} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.701044] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1164.701044] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1164.701200] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1164.701322] env[61867]: INFO nova.compute.manager [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1164.701552] env[61867]: DEBUG oslo.service.loopingcall [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1164.701775] env[61867]: DEBUG nova.compute.manager [-] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1164.701868] env[61867]: DEBUG nova.network.neutron [-] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1164.957162] env[61867]: DEBUG nova.compute.manager [req-ecfc01a0-e9e1-4afa-a4f3-0687fa4c8bd8 req-5553a9b2-8400-4d8d-99bd-f52e56dcc3a9 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Received event network-vif-deleted-5bc71c96-38d5-4cd8-92d4-46e8713e8389 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1164.957352] env[61867]: INFO nova.compute.manager [req-ecfc01a0-e9e1-4afa-a4f3-0687fa4c8bd8 req-5553a9b2-8400-4d8d-99bd-f52e56dcc3a9 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Neutron deleted interface 5bc71c96-38d5-4cd8-92d4-46e8713e8389; detaching it from the instance and deleting it from the info cache [ 1164.957578] env[61867]: DEBUG nova.network.neutron [req-ecfc01a0-e9e1-4afa-a4f3-0687fa4c8bd8 req-5553a9b2-8400-4d8d-99bd-f52e56dcc3a9 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.058872] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.059169] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.059385] env[61867]: DEBUG nova.objects.instance [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lazy-loading 'pci_requests' on Instance uuid 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1165.435791] env[61867]: DEBUG nova.network.neutron [-] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.460446] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6fb81028-3bae-43d5-abd1-c65bd70ab9ca {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.470497] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb69be07-0840-4009-bf51-223cd55f9e46 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.492377] env[61867]: DEBUG nova.compute.manager [req-ecfc01a0-e9e1-4afa-a4f3-0687fa4c8bd8 req-5553a9b2-8400-4d8d-99bd-f52e56dcc3a9 service nova] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Detach interface failed, port_id=5bc71c96-38d5-4cd8-92d4-46e8713e8389, reason: Instance 6fa6c260-9c20-4808-9dd7-946a27651681 could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1165.562547] env[61867]: DEBUG nova.objects.instance [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lazy-loading 'numa_topology' on Instance uuid 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1165.938917] env[61867]: INFO nova.compute.manager [-] [instance: 6fa6c260-9c20-4808-9dd7-946a27651681] Took 1.24 seconds to deallocate network for instance. [ 1166.065302] env[61867]: INFO nova.compute.claims [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1166.446540] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1167.113441] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-842fb228-4e24-4e1a-ab4f-49e7dc075699 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.120518] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d11f6d-c2fd-4b70-bccb-1db019164b86 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.150305] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f30ce969-422d-4986-82c7-4748c6f4252d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.157021] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff7a8288-9ddb-4f72-9dcd-6685414a1752 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.169482] env[61867]: DEBUG nova.compute.provider_tree [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1167.672380] env[61867]: DEBUG nova.scheduler.client.report [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1168.177996] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.119s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.180331] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.734s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.180571] env[61867]: DEBUG nova.objects.instance [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lazy-loading 'resources' on Instance uuid 6fa6c260-9c20-4808-9dd7-946a27651681 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.208523] env[61867]: INFO nova.network.neutron [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Updating port 5c4e4ae5-fbad-4237-ae18-41bbf1446999 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1168.721696] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6656161-215c-4837-865a-91dc4fe8fae8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.729305] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b43bf63b-a609-4d5f-8007-f5d94e44fc42 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.759292] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e01d30-1018-4aa1-9af8-8e32ccacc8ef {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.766095] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0787af48-beb5-4253-b82a-ec020e4517d2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.778756] env[61867]: DEBUG nova.compute.provider_tree [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1169.282216] env[61867]: DEBUG nova.scheduler.client.report [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1169.579230] env[61867]: DEBUG nova.compute.manager [req-8427b79f-6dc8-4aa4-a3ca-5d6dad4171d2 req-6e44a22e-74fd-4431-af86-532385f648f2 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Received event network-vif-plugged-5c4e4ae5-fbad-4237-ae18-41bbf1446999 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1169.579470] env[61867]: DEBUG oslo_concurrency.lockutils [req-8427b79f-6dc8-4aa4-a3ca-5d6dad4171d2 req-6e44a22e-74fd-4431-af86-532385f648f2 service nova] Acquiring lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.579680] env[61867]: DEBUG oslo_concurrency.lockutils [req-8427b79f-6dc8-4aa4-a3ca-5d6dad4171d2 req-6e44a22e-74fd-4431-af86-532385f648f2 service nova] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.579849] env[61867]: DEBUG oslo_concurrency.lockutils [req-8427b79f-6dc8-4aa4-a3ca-5d6dad4171d2 req-6e44a22e-74fd-4431-af86-532385f648f2 service nova] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.580076] env[61867]: DEBUG nova.compute.manager [req-8427b79f-6dc8-4aa4-a3ca-5d6dad4171d2 req-6e44a22e-74fd-4431-af86-532385f648f2 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] No waiting events found dispatching network-vif-plugged-5c4e4ae5-fbad-4237-ae18-41bbf1446999 {{(pid=61867) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1169.580288] env[61867]: WARNING nova.compute.manager [req-8427b79f-6dc8-4aa4-a3ca-5d6dad4171d2 req-6e44a22e-74fd-4431-af86-532385f648f2 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Received unexpected event network-vif-plugged-5c4e4ae5-fbad-4237-ae18-41bbf1446999 for instance with vm_state shelved_offloaded and task_state spawning. [ 1169.663329] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1169.663570] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1169.663710] env[61867]: DEBUG nova.network.neutron [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Building network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1169.787448] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.607s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.809121] env[61867]: INFO nova.scheduler.client.report [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Deleted allocations for instance 6fa6c260-9c20-4808-9dd7-946a27651681 [ 1170.316414] env[61867]: DEBUG oslo_concurrency.lockutils [None req-02a5e134-84df-473e-b848-69d9e96fab97 tempest-ServersNegativeTestJSON-1359525904 tempest-ServersNegativeTestJSON-1359525904-project-member] Lock "6fa6c260-9c20-4808-9dd7-946a27651681" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.727s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.370412] env[61867]: DEBUG nova.network.neutron [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Updating instance_info_cache with network_info: [{"id": "5c4e4ae5-fbad-4237-ae18-41bbf1446999", "address": "fa:16:3e:21:31:aa", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c4e4ae5-fb", "ovs_interfaceid": "5c4e4ae5-fbad-4237-ae18-41bbf1446999", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1170.873398] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1170.899159] env[61867]: DEBUG nova.virt.hardware [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T12:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='36ebd2b82249965f4edec1496c49a3ec',container_format='bare',created_at=2024-10-12T12:54:20Z,direct_url=,disk_format='vmdk',id=e53f4285-1fdc-4f6b-b607-bfb1722ced7f,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-825062042-shelved',owner='250e5d2fa5cb4a8a8bd0a6fdf9e50a3d',properties=ImageMetaProps,protected=,size=31666688,status='active',tags=,updated_at=2024-10-12T12:54:35Z,virtual_size=,visibility=), allow threads: False {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1170.899422] env[61867]: DEBUG nova.virt.hardware [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1170.899588] env[61867]: DEBUG nova.virt.hardware [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image limits 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1170.899774] env[61867]: DEBUG nova.virt.hardware [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Flavor pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1170.899923] env[61867]: DEBUG nova.virt.hardware [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Image pref 0:0:0 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1170.900085] env[61867]: DEBUG nova.virt.hardware [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61867) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1170.900305] env[61867]: DEBUG nova.virt.hardware [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1170.900471] env[61867]: DEBUG nova.virt.hardware [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1170.900643] env[61867]: DEBUG nova.virt.hardware [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Got 1 possible topologies {{(pid=61867) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1170.900808] env[61867]: DEBUG nova.virt.hardware [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1170.900984] env[61867]: DEBUG nova.virt.hardware [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61867) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1170.901850] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d095d867-16e2-45cb-8fa8-750551ac48da {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.909982] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d50269-1401-4edf-b2d0-8307fd9470e1 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.924522] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:31:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '399f3826-705c-45f7-9fe0-3a08a945151a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5c4e4ae5-fbad-4237-ae18-41bbf1446999', 'vif_model': 'vmxnet3'}] {{(pid=61867) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1170.931901] env[61867]: DEBUG oslo.service.loopingcall [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1170.932128] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Creating VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1170.932370] env[61867]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0c063212-0278-4d42-8dd5-e63d60f4cc6a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.951667] env[61867]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1170.951667] env[61867]: value = "task-1277265" [ 1170.951667] env[61867]: _type = "Task" [ 1170.951667] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.959728] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277265, 'name': CreateVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.461954] env[61867]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277265, 'name': CreateVM_Task, 'duration_secs': 0.301862} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.462248] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Created VM on the ESX host {{(pid=61867) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1171.462833] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e53f4285-1fdc-4f6b-b607-bfb1722ced7f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1171.463027] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e53f4285-1fdc-4f6b-b607-bfb1722ced7f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.463409] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e53f4285-1fdc-4f6b-b607-bfb1722ced7f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1171.463666] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5485aaf4-2cb0-428a-a219-8d1f222c911f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.468454] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1171.468454] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52775828-3591-b5c2-6703-d17a56659a89" [ 1171.468454] env[61867]: _type = "Task" [ 1171.468454] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.476880] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52775828-3591-b5c2-6703-d17a56659a89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.604475] env[61867]: DEBUG nova.compute.manager [req-a898f158-b04b-47a7-821d-3ef5391e6976 req-6ab6f2e9-609b-4862-89c1-2ef8f5691595 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Received event network-changed-5c4e4ae5-fbad-4237-ae18-41bbf1446999 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1171.604688] env[61867]: DEBUG nova.compute.manager [req-a898f158-b04b-47a7-821d-3ef5391e6976 req-6ab6f2e9-609b-4862-89c1-2ef8f5691595 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Refreshing instance network info cache due to event network-changed-5c4e4ae5-fbad-4237-ae18-41bbf1446999. {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1171.604908] env[61867]: DEBUG oslo_concurrency.lockutils [req-a898f158-b04b-47a7-821d-3ef5391e6976 req-6ab6f2e9-609b-4862-89c1-2ef8f5691595 service nova] Acquiring lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1171.605341] env[61867]: DEBUG oslo_concurrency.lockutils [req-a898f158-b04b-47a7-821d-3ef5391e6976 req-6ab6f2e9-609b-4862-89c1-2ef8f5691595 service nova] Acquired lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.605704] env[61867]: DEBUG nova.network.neutron [req-a898f158-b04b-47a7-821d-3ef5391e6976 req-6ab6f2e9-609b-4862-89c1-2ef8f5691595 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Refreshing network info cache for port 5c4e4ae5-fbad-4237-ae18-41bbf1446999 {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1171.979572] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e53f4285-1fdc-4f6b-b607-bfb1722ced7f" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1171.979834] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Processing image e53f4285-1fdc-4f6b-b607-bfb1722ced7f {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1171.980386] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e53f4285-1fdc-4f6b-b607-bfb1722ced7f/e53f4285-1fdc-4f6b-b607-bfb1722ced7f.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1171.980386] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e53f4285-1fdc-4f6b-b607-bfb1722ced7f/e53f4285-1fdc-4f6b-b607-bfb1722ced7f.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.980504] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1171.980664] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e16ca30a-2be8-42d2-8cc3-4b4e89b92b1d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.990512] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1171.990708] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61867) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1171.991453] env[61867]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60640f4b-be11-4976-b0c4-41fe47106e1a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.997070] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1171.997070] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f4c9e7-8a37-4cd1-51e0-d3c86a639e62" [ 1171.997070] env[61867]: _type = "Task" [ 1171.997070] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.005012] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]52f4c9e7-8a37-4cd1-51e0-d3c86a639e62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.444098] env[61867]: DEBUG nova.network.neutron [req-a898f158-b04b-47a7-821d-3ef5391e6976 req-6ab6f2e9-609b-4862-89c1-2ef8f5691595 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Updated VIF entry in instance network info cache for port 5c4e4ae5-fbad-4237-ae18-41bbf1446999. {{(pid=61867) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1172.444667] env[61867]: DEBUG nova.network.neutron [req-a898f158-b04b-47a7-821d-3ef5391e6976 req-6ab6f2e9-609b-4862-89c1-2ef8f5691595 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Updating instance_info_cache with network_info: [{"id": "5c4e4ae5-fbad-4237-ae18-41bbf1446999", "address": "fa:16:3e:21:31:aa", "network": {"id": "e2b3fc86-3698-48e2-ab7c-7da8287ae646", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522446915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "250e5d2fa5cb4a8a8bd0a6fdf9e50a3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c4e4ae5-fb", "ovs_interfaceid": "5c4e4ae5-fbad-4237-ae18-41bbf1446999", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.508015] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Preparing fetch location {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1172.508380] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Fetch image to [datastore2] OSTACK_IMG_9f2a2893-f946-4696-92b8-3f29767e7051/OSTACK_IMG_9f2a2893-f946-4696-92b8-3f29767e7051.vmdk {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1172.508482] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Downloading stream optimized image e53f4285-1fdc-4f6b-b607-bfb1722ced7f to [datastore2] OSTACK_IMG_9f2a2893-f946-4696-92b8-3f29767e7051/OSTACK_IMG_9f2a2893-f946-4696-92b8-3f29767e7051.vmdk on the data store datastore2 as vApp {{(pid=61867) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1172.508656] env[61867]: DEBUG nova.virt.vmwareapi.images [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Downloading image file data e53f4285-1fdc-4f6b-b607-bfb1722ced7f to the ESX as VM named 'OSTACK_IMG_9f2a2893-f946-4696-92b8-3f29767e7051' {{(pid=61867) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1172.585211] env[61867]: DEBUG oslo_vmware.rw_handles [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1172.585211] env[61867]: value = "resgroup-9" [ 1172.585211] env[61867]: _type = "ResourcePool" [ 1172.585211] env[61867]: }. {{(pid=61867) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1172.585880] env[61867]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-258ce7d7-7382-402c-88ef-e29623cdfdf0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.607827] env[61867]: DEBUG oslo_vmware.rw_handles [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lease: (returnval){ [ 1172.607827] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5279e314-a309-4101-0432-b0b6aeb25b55" [ 1172.607827] env[61867]: _type = "HttpNfcLease" [ 1172.607827] env[61867]: } obtained for vApp import into resource pool (val){ [ 1172.607827] env[61867]: value = "resgroup-9" [ 1172.607827] env[61867]: _type = "ResourcePool" [ 1172.607827] env[61867]: }. {{(pid=61867) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1172.608171] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the lease: (returnval){ [ 1172.608171] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5279e314-a309-4101-0432-b0b6aeb25b55" [ 1172.608171] env[61867]: _type = "HttpNfcLease" [ 1172.608171] env[61867]: } to be ready. {{(pid=61867) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1172.615091] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1172.615091] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5279e314-a309-4101-0432-b0b6aeb25b55" [ 1172.615091] env[61867]: _type = "HttpNfcLease" [ 1172.615091] env[61867]: } is initializing. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1172.947948] env[61867]: DEBUG oslo_concurrency.lockutils [req-a898f158-b04b-47a7-821d-3ef5391e6976 req-6ab6f2e9-609b-4862-89c1-2ef8f5691595 service nova] Releasing lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1173.116900] env[61867]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1173.116900] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5279e314-a309-4101-0432-b0b6aeb25b55" [ 1173.116900] env[61867]: _type = "HttpNfcLease" [ 1173.116900] env[61867]: } is ready. {{(pid=61867) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1173.117239] env[61867]: DEBUG oslo_vmware.rw_handles [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1173.117239] env[61867]: value = "session[52bf9b2b-af16-ecb2-caf2-6191bbdff809]5279e314-a309-4101-0432-b0b6aeb25b55" [ 1173.117239] env[61867]: _type = "HttpNfcLease" [ 1173.117239] env[61867]: }. {{(pid=61867) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1173.117950] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a177bfa9-1682-4b84-870d-5247a3b0c072 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.126081] env[61867]: DEBUG oslo_vmware.rw_handles [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526f5b75-5b9e-b1eb-c169-6db0fa651832/disk-0.vmdk from lease info. {{(pid=61867) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1173.126334] env[61867]: DEBUG oslo_vmware.rw_handles [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Creating HTTP connection to write to file with size = 31666688 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526f5b75-5b9e-b1eb-c169-6db0fa651832/disk-0.vmdk. {{(pid=61867) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1173.193186] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-380deca7-b94e-42cc-a5cc-96cb3813b9aa {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.286202] env[61867]: DEBUG oslo_vmware.rw_handles [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Completed reading data from the image iterator. {{(pid=61867) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1174.286550] env[61867]: DEBUG oslo_vmware.rw_handles [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526f5b75-5b9e-b1eb-c169-6db0fa651832/disk-0.vmdk. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1174.287502] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea93bc29-1427-4445-88a9-c8ffd82372b8 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.294930] env[61867]: DEBUG oslo_vmware.rw_handles [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526f5b75-5b9e-b1eb-c169-6db0fa651832/disk-0.vmdk is in state: ready. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1174.295155] env[61867]: DEBUG oslo_vmware.rw_handles [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526f5b75-5b9e-b1eb-c169-6db0fa651832/disk-0.vmdk. {{(pid=61867) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1174.295457] env[61867]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-75a205a3-8b9f-45f9-b76c-f1df8c42dbab {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.712496] env[61867]: DEBUG oslo_vmware.rw_handles [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526f5b75-5b9e-b1eb-c169-6db0fa651832/disk-0.vmdk. {{(pid=61867) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1174.712737] env[61867]: INFO nova.virt.vmwareapi.images [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Downloaded image file data e53f4285-1fdc-4f6b-b607-bfb1722ced7f [ 1174.713664] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101d66b2-0787-4261-8a44-2ab656ae8a45 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.729730] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7ee1b40a-2d2a-4421-9b7a-1de20a1f2a8a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.756136] env[61867]: INFO nova.virt.vmwareapi.images [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] The imported VM was unregistered [ 1174.758512] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Caching image {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1174.758750] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Creating directory with path [datastore2] devstack-image-cache_base/e53f4285-1fdc-4f6b-b607-bfb1722ced7f {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1174.759045] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c16772d-102a-4917-882f-b9ea95d14ce5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.773015] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Created directory with path [datastore2] devstack-image-cache_base/e53f4285-1fdc-4f6b-b607-bfb1722ced7f {{(pid=61867) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1174.773296] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_9f2a2893-f946-4696-92b8-3f29767e7051/OSTACK_IMG_9f2a2893-f946-4696-92b8-3f29767e7051.vmdk to [datastore2] devstack-image-cache_base/e53f4285-1fdc-4f6b-b607-bfb1722ced7f/e53f4285-1fdc-4f6b-b607-bfb1722ced7f.vmdk. {{(pid=61867) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1174.773580] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-a49d92a5-1bf9-451c-856f-46b7fba878b5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.781900] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1174.781900] env[61867]: value = "task-1277268" [ 1174.781900] env[61867]: _type = "Task" [ 1174.781900] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.790526] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277268, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.292774] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277268, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.792849] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277268, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.294418] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277268, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.795209] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277268, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.294771] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277268, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.165505} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.295191] env[61867]: INFO nova.virt.vmwareapi.ds_util [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_9f2a2893-f946-4696-92b8-3f29767e7051/OSTACK_IMG_9f2a2893-f946-4696-92b8-3f29767e7051.vmdk to [datastore2] devstack-image-cache_base/e53f4285-1fdc-4f6b-b607-bfb1722ced7f/e53f4285-1fdc-4f6b-b607-bfb1722ced7f.vmdk. [ 1177.295262] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Cleaning up location [datastore2] OSTACK_IMG_9f2a2893-f946-4696-92b8-3f29767e7051 {{(pid=61867) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1177.295416] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_9f2a2893-f946-4696-92b8-3f29767e7051 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1177.295680] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3fe7d6b6-3d5b-4be9-96ea-daef23c7b2e2 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.302607] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1177.302607] env[61867]: value = "task-1277269" [ 1177.302607] env[61867]: _type = "Task" [ 1177.302607] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.310229] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277269, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.812762] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277269, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.032984} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.813027] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1177.813213] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e53f4285-1fdc-4f6b-b607-bfb1722ced7f/e53f4285-1fdc-4f6b-b607-bfb1722ced7f.vmdk" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.813460] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e53f4285-1fdc-4f6b-b607-bfb1722ced7f/e53f4285-1fdc-4f6b-b607-bfb1722ced7f.vmdk to [datastore2] 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66/214bb1c3-3d79-4ad8-9f1a-aad95dc83d66.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1177.813713] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-806f350e-3f7d-41a1-bfb1-a676b0d1926f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.820604] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1177.820604] env[61867]: value = "task-1277270" [ 1177.820604] env[61867]: _type = "Task" [ 1177.820604] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.828188] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277270, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.332283] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277270, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.832885] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277270, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.333690] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277270, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.834600] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277270, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.334418] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277270, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.120728} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.334719] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e53f4285-1fdc-4f6b-b607-bfb1722ced7f/e53f4285-1fdc-4f6b-b607-bfb1722ced7f.vmdk to [datastore2] 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66/214bb1c3-3d79-4ad8-9f1a-aad95dc83d66.vmdk {{(pid=61867) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1180.335537] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61fc71c3-f3f0-4e38-87f9-5ef6c3483d8e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.358161] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66/214bb1c3-3d79-4ad8-9f1a-aad95dc83d66.vmdk or device None with type streamOptimized {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1180.358416] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2d81087-6828-4918-9534-9bb75bf276e0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.378079] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1180.378079] env[61867]: value = "task-1277271" [ 1180.378079] env[61867]: _type = "Task" [ 1180.378079] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.385992] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277271, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.888602] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277271, 'name': ReconfigVM_Task, 'duration_secs': 0.271434} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.888903] env[61867]: DEBUG nova.virt.vmwareapi.volumeops [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66/214bb1c3-3d79-4ad8-9f1a-aad95dc83d66.vmdk or device None with type streamOptimized {{(pid=61867) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1180.889542] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2d32d86f-e0c5-4f9a-be8a-f187b8ee412f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.895878] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1180.895878] env[61867]: value = "task-1277272" [ 1180.895878] env[61867]: _type = "Task" [ 1180.895878] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.903369] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277272, 'name': Rename_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.406687] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277272, 'name': Rename_Task, 'duration_secs': 0.131613} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.407011] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Powering on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1181.407211] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b02e191-fa94-4bc0-a9e9-9135862114fb {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.414147] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1181.414147] env[61867]: value = "task-1277273" [ 1181.414147] env[61867]: _type = "Task" [ 1181.414147] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.421739] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277273, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.924750] env[61867]: DEBUG oslo_vmware.api [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277273, 'name': PowerOnVM_Task, 'duration_secs': 0.461147} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.925040] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Powered on the VM {{(pid=61867) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1181.957532] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1182.029594] env[61867]: DEBUG nova.compute.manager [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Checking state {{(pid=61867) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1182.030554] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee39d2a7-e470-4af1-8eda-3ed0ee19f2df {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.549354] env[61867]: DEBUG oslo_concurrency.lockutils [None req-9efd5a39-d596-49e7-b522-c7ecb1597ede tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.515s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1182.957543] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1183.319054] env[61867]: DEBUG oslo_concurrency.lockutils [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.319054] env[61867]: DEBUG oslo_concurrency.lockutils [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.319054] env[61867]: DEBUG oslo_concurrency.lockutils [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.319054] env[61867]: DEBUG oslo_concurrency.lockutils [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.319345] env[61867]: DEBUG oslo_concurrency.lockutils [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.321328] env[61867]: INFO nova.compute.manager [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Terminating instance [ 1183.323108] env[61867]: DEBUG nova.compute.manager [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Start destroying the instance on the hypervisor. {{(pid=61867) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1183.323317] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Destroying instance {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1183.324150] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-257bcce9-de83-4a89-9161-591851076411 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.332695] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Powering off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1183.332920] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-575ebad1-c2eb-436f-a2f3-fe6fab9fd330 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.339602] env[61867]: DEBUG oslo_vmware.api [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1183.339602] env[61867]: value = "task-1277274" [ 1183.339602] env[61867]: _type = "Task" [ 1183.339602] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.347318] env[61867]: DEBUG oslo_vmware.api [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277274, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.850144] env[61867]: DEBUG oslo_vmware.api [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277274, 'name': PowerOffVM_Task, 'duration_secs': 0.185717} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.850546] env[61867]: DEBUG nova.virt.vmwareapi.vm_util [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Powered off the VM {{(pid=61867) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1183.850588] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Unregistering the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1183.850818] env[61867]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5e991aeb-9e95-4625-81df-7b154bc0cd58 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.911959] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Unregistered the VM {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1183.912185] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Deleting contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1183.912394] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Deleting the datastore file [datastore2] 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66 {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1183.912668] env[61867]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5b0859fc-1155-41b3-af4d-d6e23794e20a {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.922136] env[61867]: DEBUG oslo_vmware.api [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for the task: (returnval){ [ 1183.922136] env[61867]: value = "task-1277276" [ 1183.922136] env[61867]: _type = "Task" [ 1183.922136] env[61867]: } to complete. {{(pid=61867) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.930498] env[61867]: DEBUG oslo_vmware.api [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277276, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.957162] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1184.431771] env[61867]: DEBUG oslo_vmware.api [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Task: {'id': task-1277276, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146442} completed successfully. {{(pid=61867) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.432063] env[61867]: DEBUG nova.virt.vmwareapi.ds_util [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Deleted the datastore file {{(pid=61867) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1184.432243] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Deleted contents of the VM from datastore datastore2 {{(pid=61867) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1184.432450] env[61867]: DEBUG nova.virt.vmwareapi.vmops [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Instance destroyed {{(pid=61867) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1184.432643] env[61867]: INFO nova.compute.manager [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1184.432888] env[61867]: DEBUG oslo.service.loopingcall [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61867) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1184.433097] env[61867]: DEBUG nova.compute.manager [-] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Deallocating network for instance {{(pid=61867) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1184.433193] env[61867]: DEBUG nova.network.neutron [-] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] deallocate_for_instance() {{(pid=61867) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1184.879435] env[61867]: DEBUG nova.compute.manager [req-3028bee5-b586-4531-8cd6-a1626c108014 req-be97d677-61f9-470d-8cc1-f09f32119238 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Received event network-vif-deleted-5c4e4ae5-fbad-4237-ae18-41bbf1446999 {{(pid=61867) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1184.879725] env[61867]: INFO nova.compute.manager [req-3028bee5-b586-4531-8cd6-a1626c108014 req-be97d677-61f9-470d-8cc1-f09f32119238 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Neutron deleted interface 5c4e4ae5-fbad-4237-ae18-41bbf1446999; detaching it from the instance and deleting it from the info cache [ 1184.879890] env[61867]: DEBUG nova.network.neutron [req-3028bee5-b586-4531-8cd6-a1626c108014 req-be97d677-61f9-470d-8cc1-f09f32119238 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.956811] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.347049] env[61867]: DEBUG nova.network.neutron [-] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1185.383071] env[61867]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3737d1b4-d375-470d-b0d7-03b816eff2ca {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.393368] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7efcee69-3e08-44ff-9d7f-6d03c1dfd56b {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.418907] env[61867]: DEBUG nova.compute.manager [req-3028bee5-b586-4531-8cd6-a1626c108014 req-be97d677-61f9-470d-8cc1-f09f32119238 service nova] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Detach interface failed, port_id=5c4e4ae5-fbad-4237-ae18-41bbf1446999, reason: Instance 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66 could not be found. {{(pid=61867) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1185.459881] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.460271] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.460527] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.460659] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61867) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1185.461857] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c77c9f-6237-4b38-8091-e66429e300b4 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.470086] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b808456-ce79-4b79-b1be-370a4e670d3f {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.486374] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf7d7140-8b14-416b-9369-8cdc369e287e {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.508477] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4caee41-2246-4e0e-9434-2b23db684190 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.538391] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180898MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61867) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1185.538530] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.538697] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.849522] env[61867]: INFO nova.compute.manager [-] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Took 1.42 seconds to deallocate network for instance. [ 1186.356074] env[61867]: DEBUG oslo_concurrency.lockutils [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.564677] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Instance 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61867) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1186.564901] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1186.565060] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61867) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1186.590913] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5bb847d-2a4c-4150-9ed7-6e90b55da6e7 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.598674] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f59e446e-e29d-4dcc-b284-615a3a0f0297 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.627546] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c4fb307-be45-4225-bec0-a85beaf9bae0 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.634142] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d91627-33d2-4989-a888-eae23edc7b5d {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.647960] env[61867]: DEBUG nova.compute.provider_tree [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1187.151199] env[61867]: DEBUG nova.scheduler.client.report [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1187.656497] env[61867]: DEBUG nova.compute.resource_tracker [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61867) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1187.656887] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.118s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.657035] env[61867]: DEBUG oslo_concurrency.lockutils [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.301s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.657260] env[61867]: DEBUG nova.objects.instance [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lazy-loading 'resources' on Instance uuid 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1188.186700] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9199c09-8e55-4177-8262-c878c5fd1995 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.194258] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86038ea3-513a-4703-9af9-dbbca8161952 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.222882] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c2019f-079c-445d-88bf-3a891a3e38d5 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.229477] env[61867]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6c6183-ef54-40be-9176-2e642a15fc84 {{(pid=61867) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.241747] env[61867]: DEBUG nova.compute.provider_tree [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed in ProviderTree for provider: 25720271-a549-4916-abe3-e5ed9b765889 {{(pid=61867) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1188.654384] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.654610] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.654756] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Starting heal instance info cache {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1188.654874] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Rebuilding the list of instances to heal {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1188.744928] env[61867]: DEBUG nova.scheduler.client.report [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Inventory has not changed for provider 25720271-a549-4916-abe3-e5ed9b765889 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61867) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1189.173962] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquiring lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.174134] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Acquired lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.174285] env[61867]: DEBUG nova.network.neutron [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Forcefully refreshing network info cache for instance {{(pid=61867) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1189.174437] env[61867]: DEBUG nova.objects.instance [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Lazy-loading 'info_cache' on Instance uuid 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66 {{(pid=61867) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1189.249635] env[61867]: DEBUG oslo_concurrency.lockutils [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.593s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.272216] env[61867]: INFO nova.scheduler.client.report [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Deleted allocations for instance 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66 [ 1189.779491] env[61867]: DEBUG oslo_concurrency.lockutils [None req-88944809-df69-4d02-b39f-6dfc7e4bba63 tempest-ServerActionsTestOtherB-1406828891 tempest-ServerActionsTestOtherB-1406828891-project-member] Lock "214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.461s {{(pid=61867) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.181085] env[61867]: DEBUG nova.compute.utils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Can not refresh info_cache because instance was not found {{(pid=61867) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 1190.198522] env[61867]: DEBUG nova.network.neutron [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Instance cache missing network info. {{(pid=61867) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1190.780714] env[61867]: DEBUG nova.network.neutron [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Updating instance_info_cache with network_info: [] {{(pid=61867) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1191.283125] env[61867]: DEBUG oslo_concurrency.lockutils [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Releasing lock "refresh_cache-214bb1c3-3d79-4ad8-9f1a-aad95dc83d66" {{(pid=61867) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1191.283343] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] [instance: 214bb1c3-3d79-4ad8-9f1a-aad95dc83d66] Updated the network info_cache for instance {{(pid=61867) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1191.283561] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1191.283734] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1191.283863] env[61867]: DEBUG nova.compute.manager [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61867) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1192.957401] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.952560] env[61867]: DEBUG oslo_service.periodic_task [None req-0be69a5c-0ab7-4beb-82bd-afff65e02658 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61867) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}